/*
- * Copyright (c) 2000-2010 Apple Inc. All rights reserved.
+ * Copyright (c) 2000-2016 Apple Inc. All rights reserved.
*
* @APPLE_OSREFERENCE_LICENSE_HEADER_START@
*
* @APPLE_OSREFERENCE_LICENSE_HEADER_END@
*/
#ifdef MACH_BSD
-#include <mach_rt.h>
#include <mach_debug.h>
#include <mach_ldebug.h>
#ifdef MACH_BSD
extern void mach_kauth_cred_uthread_update(void);
+extern void throttle_lowpri_io(int);
#endif
void * find_user_regs(thread_t);
unsigned int get_msr_rbits(void);
-extern void throttle_lowpri_io(int);
-
/*
* thread_userstack:
*
thread_state_t tstate,
__unused unsigned int count,
mach_vm_offset_t *user_stack,
- int *customstack
+ int *customstack,
+ __unused boolean_t is64bit
)
{
if (customstack)
return (KERN_SUCCESS);
}
+/*
+ * thread_userstackdefault:
+ *
+ * Return the default stack location for the
+ * thread, if otherwise unknown.
+ */
+kern_return_t
+thread_userstackdefault(
+ mach_vm_offset_t *default_user_stack,
+ boolean_t is64bit)
+{
+ if (is64bit) {
+ *default_user_stack = VM_USRSTACK64;
+ } else {
+ *default_user_stack = VM_USRSTACK32;
+ }
+ return (KERN_SUCCESS);
+}
kern_return_t
thread_entrypoint(
extern long fuword(vm_offset_t);
-
-
+__attribute__((noreturn))
void
machdep_syscall(x86_saved_state_t *state)
{
int args[machdep_call_count];
int trapno;
int nargs;
- machdep_call_t *entry;
+ const machdep_call_t *entry;
x86_saved_state32_t *regs;
assert(is_saved_state32(state));
default:
panic("machdep_syscall: too many args");
}
- if (current_thread()->funnel_lock)
- (void) thread_funnel_set(current_thread()->funnel_lock, FALSE);
DEBUG_KPRINT_SYSCALL_MDEP("machdep_syscall: retval=%u\n", regs->eax);
- throttle_lowpri_io(TRUE);
+#if DEBUG || DEVELOPMENT
+ kern_allocation_name_t
+ prior __assert_only = thread_get_kernel_state(current_thread())->allocation_name;
+ assertf(prior == NULL, "thread_set_allocation_name(\"%s\") not cleared", kern_allocation_get_name(prior));
+#endif /* DEBUG || DEVELOPMENT */
+
+ throttle_lowpri_io(1);
thread_exception_return();
/* NOTREACHED */
}
-
+__attribute__((noreturn))
void
machdep_syscall64(x86_saved_state_t *state)
{
int trapno;
- machdep_call_t *entry;
+ const machdep_call_t *entry;
x86_saved_state64_t *regs;
assert(is_saved_state64(state));
case 1:
regs->rax = (*entry->routine.args64_1)(regs->rdi);
break;
+ case 2:
+ regs->rax = (*entry->routine.args64_2)(regs->rdi, regs->rsi);
+ break;
default:
panic("machdep_syscall64: too many args");
}
- if (current_thread()->funnel_lock)
- (void) thread_funnel_set(current_thread()->funnel_lock, FALSE);
DEBUG_KPRINT_SYSCALL_MDEP("machdep_syscall: retval=%llu\n", regs->rax);
- throttle_lowpri_io(TRUE);
+#if DEBUG || DEVELOPMENT
+ kern_allocation_name_t
+ prior __assert_only = thread_get_kernel_state(current_thread())->allocation_name;
+ assertf(prior == NULL, "thread_set_allocation_name(\"%s\") not cleared", kern_allocation_get_name(prior));
+#endif /* DEBUG || DEVELOPMENT */
+
+ throttle_lowpri_io(1);
thread_exception_return();
/* NOTREACHED */
};
static kern_return_t
-mach_call_arg_munger32(uint32_t sp, int nargs, int call_number, struct mach_call_args *args);
+mach_call_arg_munger32(uint32_t sp, struct mach_call_args *args, const mach_trap_t *trapp);
static kern_return_t
-mach_call_arg_munger32(uint32_t sp, int nargs, int call_number, struct mach_call_args *args)
+mach_call_arg_munger32(uint32_t sp, struct mach_call_args *args, const mach_trap_t *trapp)
{
- unsigned int args32[9];
-
- if (copyin((user_addr_t)(sp + sizeof(int)), (char *)args32, nargs * sizeof (int)))
+ if (copyin((user_addr_t)(sp + sizeof(int)), (char *)args, trapp->mach_trap_u32_words * sizeof (int)))
return KERN_INVALID_ARGUMENT;
-
- switch (nargs) {
- case 9: args->arg9 = args32[8];
- case 8: args->arg8 = args32[7];
- case 7: args->arg7 = args32[6];
- case 6: args->arg6 = args32[5];
- case 5: args->arg5 = args32[4];
- case 4: args->arg4 = args32[3];
- case 3: args->arg3 = args32[2];
- case 2: args->arg2 = args32[1];
- case 1: args->arg1 = args32[0];
- }
- if (call_number == 90) {
- /* munge_l for mach_wait_until_trap() */
- args->arg1 = (((uint64_t)(args32[0])) | ((((uint64_t)(args32[1]))<<32)));
- }
- if (call_number == 93) {
- /* munge_wl for mk_timer_arm_trap() */
- args->arg2 = (((uint64_t)(args32[1])) | ((((uint64_t)(args32[2]))<<32)));
- }
-
+#if CONFIG_REQUIRES_U32_MUNGING
+ trapp->mach_trap_arg_munge32(args);
+#else
+#error U32 mach traps on x86_64 kernel requires munging
+#endif
return KERN_SUCCESS;
}
extern const char *mach_syscall_name_table[];
+__attribute__((noreturn))
void
mach_call_munger(x86_saved_state_t *state)
{
struct mach_call_args args = { 0, 0, 0, 0, 0, 0, 0, 0, 0 };
x86_saved_state32_t *regs;
+ struct uthread *ut = get_bsdthread_info(current_thread());
+ uthread_reset_proc_refcount(ut);
+
assert(is_saved_state32(state));
regs = saved_state32(state);
argc = mach_trap_table[call_number].mach_trap_arg_count;
if (argc) {
- retval = mach_call_arg_munger32(regs->uesp, argc, call_number, &args);
+ retval = mach_call_arg_munger32(regs->uesp, &args, &mach_trap_table[call_number]);
if (retval != KERN_SUCCESS) {
regs->eax = retval;
#ifdef MACH_BSD
mach_kauth_cred_uthread_update();
#endif
- KERNEL_DEBUG_CONSTANT(MACHDBG_CODE(DBG_MACH_EXCP_SC, (call_number)) | DBG_FUNC_START,
- args.arg1, args.arg2, args.arg3, args.arg4, 0);
+
+ KERNEL_DEBUG_CONSTANT_IST(KDEBUG_TRACE,
+ MACHDBG_CODE(DBG_MACH_EXCP_SC, (call_number)) | DBG_FUNC_START,
+ args.arg1, args.arg2, args.arg3, args.arg4, 0);
retval = mach_call(&args);
DEBUG_KPRINT_SYSCALL_MACH("mach_call_munger: retval=0x%x\n", retval);
- KERNEL_DEBUG_CONSTANT(MACHDBG_CODE(DBG_MACH_EXCP_SC,(call_number)) | DBG_FUNC_END,
- retval, 0, 0, 0, 0);
+ KERNEL_DEBUG_CONSTANT_IST(KDEBUG_TRACE,
+ MACHDBG_CODE(DBG_MACH_EXCP_SC,(call_number)) | DBG_FUNC_END,
+ retval, 0, 0, 0, 0);
+
regs->eax = retval;
- throttle_lowpri_io(TRUE);
+#if DEBUG || DEVELOPMENT
+ kern_allocation_name_t
+ prior __assert_only = thread_get_kernel_state(current_thread())->allocation_name;
+ assertf(prior == NULL, "thread_set_allocation_name(\"%s\") not cleared", kern_allocation_get_name(prior));
+#endif /* DEBUG || DEVELOPMENT */
+
+ throttle_lowpri_io(1);
+
+#if PROC_REF_DEBUG
+ if (__improbable(uthread_get_proc_refcount(ut) != 0)) {
+ panic("system call returned with uu_proc_refcount != 0");
+ }
+#endif
thread_exception_return();
/* NOTREACHED */
__private_extern__ void mach_call_munger64(x86_saved_state_t *regs);
+__attribute__((noreturn))
void
mach_call_munger64(x86_saved_state_t *state)
{
int call_number;
int argc;
mach_call_t mach_call;
+ struct mach_call_args args = { 0, 0, 0, 0, 0, 0, 0, 0, 0 };
x86_saved_state64_t *regs;
+ struct uthread *ut = get_bsdthread_info(current_thread());
+ uthread_reset_proc_refcount(ut);
+
assert(is_saved_state64(state));
regs = saved_state64(state);
"mach_call_munger64: code=%d(%s)\n",
call_number, mach_syscall_name_table[call_number]);
- KERNEL_DEBUG_CONSTANT(MACHDBG_CODE(DBG_MACH_EXCP_SC,
- (call_number)) | DBG_FUNC_START,
- regs->rdi, regs->rsi,
- regs->rdx, regs->r10, 0);
+ KERNEL_DEBUG_CONSTANT_IST(KDEBUG_TRACE,
+ MACHDBG_CODE(DBG_MACH_EXCP_SC,(call_number)) | DBG_FUNC_START,
+ regs->rdi, regs->rsi, regs->rdx, regs->r10, 0);
if (call_number < 0 || call_number >= mach_trap_count) {
i386_exception(EXC_SYSCALL, regs->rax, 1);
/* NOTREACHED */
}
argc = mach_trap_table[call_number].mach_trap_arg_count;
+ if (argc) {
+ int args_in_regs = MIN(6, argc);
- if (argc > 6) {
+ memcpy(&args.arg1, ®s->rdi, args_in_regs * sizeof(syscall_arg_t));
+
+ if (argc > 6) {
int copyin_count;
- copyin_count = (argc - 6) * (int)sizeof(uint64_t);
+ assert(argc <= 9);
+ copyin_count = (argc - 6) * (int)sizeof(syscall_arg_t);
- if (copyin((user_addr_t)(regs->isf.rsp + sizeof(user_addr_t)), (char *)®s->v_arg6, copyin_count)) {
+ if (copyin((user_addr_t)(regs->isf.rsp + sizeof(user_addr_t)), (char *)&args.arg7, copyin_count)) {
regs->rax = KERN_INVALID_ARGUMENT;
- thread_exception_return();
- /* NOTREACHED */
+ thread_exception_return();
+ /* NOTREACHED */
+ }
}
}
mach_kauth_cred_uthread_update();
#endif
- regs->rax = (uint64_t)mach_call((void *)(®s->rdi));
+ regs->rax = (uint64_t)mach_call((void *)&args);
DEBUG_KPRINT_SYSCALL_MACH( "mach_call_munger64: retval=0x%llx\n", regs->rax);
- KERNEL_DEBUG_CONSTANT(MACHDBG_CODE(DBG_MACH_EXCP_SC,
- (call_number)) | DBG_FUNC_END,
- regs->rax, 0, 0, 0, 0);
+ KERNEL_DEBUG_CONSTANT_IST(KDEBUG_TRACE,
+ MACHDBG_CODE(DBG_MACH_EXCP_SC,(call_number)) | DBG_FUNC_END,
+ regs->rax, 0, 0, 0, 0);
+
+#if DEBUG || DEVELOPMENT
+ kern_allocation_name_t
+ prior __assert_only = thread_get_kernel_state(current_thread())->allocation_name;
+ assertf(prior == NULL, "thread_set_allocation_name(\"%s\") not cleared", kern_allocation_get_name(prior));
+#endif /* DEBUG || DEVELOPMENT */
- throttle_lowpri_io(TRUE);
+ throttle_lowpri_io(1);
+
+#if PROC_REF_DEBUG
+ if (__improbable(uthread_get_proc_refcount(ut) != 0)) {
+ panic("system call returned with uu_proc_refcount != 0");
+ }
+#endif
thread_exception_return();
/* NOTREACHED */
return (KERN_SUCCESS);
}
-
-
-/* XXX this should be a struct savearea so that CHUD will work better on x86 */
void *
-find_user_regs(thread_t thread)
+get_user_regs(thread_t th)
{
- pal_register_cache_state(thread, DIRTY);
- return USER_STATE(thread);
+ pal_register_cache_state(th, DIRTY);
+ return(USER_STATE(th));
}
void *
-get_user_regs(thread_t th)
+find_user_regs(thread_t thread)
{
- pal_register_cache_state(th, DIRTY);
- return(USER_STATE(th));
+ return get_user_regs(thread);
}
#if CONFIG_DTRACE