X-Git-Url: https://git.saurik.com/apple/xnu.git/blobdiff_plain/1c79356b52d46aa6b508fb032f5ae709b1f2897b..04b8595b18b1b41ac7a206e4b3d51a635f8413d7:/bsd/kern/kern_core.c diff --git a/bsd/kern/kern_core.c b/bsd/kern/kern_core.c index 7776eb45d..2bd9de059 100644 --- a/bsd/kern/kern_core.c +++ b/bsd/kern/kern_core.c @@ -1,23 +1,29 @@ /* - * Copyright (c) 2000 Apple Computer, Inc. All rights reserved. + * Copyright (c) 2000-2006 Apple Computer, Inc. All rights reserved. * - * @APPLE_LICENSE_HEADER_START@ + * @APPLE_OSREFERENCE_LICENSE_HEADER_START@ * - * The contents of this file constitute Original Code as defined in and - * are subject to the Apple Public Source License Version 1.1 (the - * "License"). You may not use this file except in compliance with the - * License. Please obtain a copy of the License at - * http://www.apple.com/publicsource and read it before using this file. + * This file contains Original Code and/or Modifications of Original Code + * as defined in and that are subject to the Apple Public Source License + * Version 2.0 (the 'License'). You may not use this file except in + * compliance with the License. The rights granted to you under the License + * may not be used to create, or enable the creation or redistribution of, + * unlawful or unlicensed copies of an Apple operating system, or to + * circumvent, violate, or enable the circumvention or violation of, any + * terms of an Apple operating system software license agreement. * - * This Original Code and all software distributed under the License are - * distributed on an "AS IS" basis, WITHOUT WARRANTY OF ANY KIND, EITHER + * Please obtain a copy of the License at + * http://www.opensource.apple.com/apsl/ and read it before using this file. + * + * The Original Code and all software distributed under the License are + * distributed on an 'AS IS' basis, WITHOUT WARRANTY OF ANY KIND, EITHER * EXPRESS OR IMPLIED, AND APPLE HEREBY DISCLAIMS ALL SUCH WARRANTIES, * INCLUDING WITHOUT LIMITATION, ANY WARRANTIES OF MERCHANTABILITY, - * FITNESS FOR A PARTICULAR PURPOSE OR NON-INFRINGEMENT. Please see the - * License for the specific language governing rights and limitations - * under the License. + * FITNESS FOR A PARTICULAR PURPOSE, QUIET ENJOYMENT OR NON-INFRINGEMENT. + * Please see the License for the specific language governing rights and + * limitations under the License. * - * @APPLE_LICENSE_HEADER_END@ + * @APPLE_OSREFERENCE_LICENSE_HEADER_END@ */ /* Copyright (c) 1991 NeXT Computer, Inc. All rights reserved. * @@ -25,9 +31,6 @@ * * This file contains machine independent code for performing core dumps. * - * HISTORY - * 16-Feb-91 Mike DeMoney (mike@next.com) - * Massaged into MI form from m68k/core.c. */ #include @@ -38,49 +41,42 @@ #include #include #include -#include -#include +#include +#include +#include #include #include -#include #include -#include +#include #include #include #include #include #include +#include #include +#include /* last */ +#include /* current_map() */ +#include /* mach_vm_region_recurse() */ +#include /* task_suspend() */ +#include /* get_task_numacts() */ + +#include typedef struct { int flavor; /* the number for this flavor */ - int count; /* count of ints in this flavor */ + mach_msg_type_number_t count; /* count of ints in this flavor */ } mythread_state_flavor_t; -#if defined (__ppc__) - -mythread_state_flavor_t thread_flavor_array[]={ - {PPC_THREAD_STATE , PPC_THREAD_STATE_COUNT}, - {PPC_FLOAT_STATE, PPC_FLOAT_STATE_COUNT}, - {PPC_EXCEPTION_STATE, PPC_EXCEPTION_STATE_COUNT} - }; -int mynum_flavors=3; -#elif defined (__i386__) +#if defined (__i386__) || defined (__x86_64__) mythread_state_flavor_t thread_flavor_array [] = { - {i386_THREAD_STATE, i386_THREAD_STATE_COUNT}, - {i386_THREAD_FPSTATE, i386_THREAD_FPSTATE_COUNT}, - {i386_THREAD_EXCEPTSTATE, i386_THREAD_EXCEPTSTATE_COUNT}, - {i386_THREAD_CTHREADSTATE, i386_THREAD_CTHREADSTATE_COUNT}, - {i386_NEW_THREAD_STATE, i386_NEW_THREAD_STATE_COUNT}, - {i386_FLOAT_STATE, i386_FLOAT_STATE_COUNT}, - {i386_ISA_PORT_MAP_STATE, i386_ISA_PORT_MAP_STATE_COUNT}, - {i386_V86_ASSIST_STATE, i386_V86_ASSIST_STATE_COUNT}, - {THREAD_SYSCALL_STATE, i386_THREAD_SYSCALL_STATE_COUNT} + {x86_THREAD_STATE, x86_THREAD_STATE_COUNT}, + {x86_FLOAT_STATE, x86_FLOAT_STATE_COUNT}, + {x86_EXCEPTION_STATE, x86_EXCEPTION_STATE_COUNT}, }; -int mynum_flavors=9; - +int mynum_flavors=3; #else #error architecture not supported #endif @@ -91,14 +87,71 @@ typedef struct { int hoffset; mythread_state_flavor_t *flavors; int tstate_size; + int flavor_count; } tir_t; -collectth_state(thread_act_t th_act, tir_t *t) +/* XXX should be static */ +void collectth_state(thread_t th_act, void *tirp); + +extern int freespace_mb(vnode_t vp); + +/* XXX not in a Mach header anywhere */ +kern_return_t thread_getstatus(register thread_t act, int flavor, + thread_state_t tstate, mach_msg_type_number_t *count); +void task_act_iterate_wth_args(task_t, void(*)(thread_t, void *), void *); +extern kern_return_t task_suspend_internal(task_t); + +static cpu_type_t process_cpu_type(proc_t proc); +static cpu_type_t process_cpu_subtype(proc_t proc); + +#ifdef SECURE_KERNEL +__XNU_PRIVATE_EXTERN int do_coredump = 0; /* default: don't dump cores */ +#else +__XNU_PRIVATE_EXTERN int do_coredump = 1; /* default: dump cores */ +#endif +__XNU_PRIVATE_EXTERN int sugid_coredump = 0; /* default: but not SGUID binaries */ + + +/* cpu_type returns only the most generic indication of the current CPU. */ +/* in a core we want to know the kind of process. */ + +static cpu_type_t +process_cpu_type(proc_t core_proc) +{ + cpu_type_t what_we_think; +#if defined (__i386__) || defined (__x86_64__) + if (IS_64BIT_PROCESS(core_proc)) { + what_we_think = CPU_TYPE_X86_64; + } else { + what_we_think = CPU_TYPE_I386; + } +#endif + return what_we_think; +} + +static cpu_type_t +process_cpu_subtype(proc_t core_proc) +{ + cpu_type_t what_we_think; +#if defined (__i386__) || defined (__x86_64__) + if (IS_64BIT_PROCESS(core_proc)) { + what_we_think = CPU_SUBTYPE_X86_64_ALL; + } else { + what_we_think = CPU_SUBTYPE_I386_ALL; + } +#endif + return what_we_think; +} + +void +collectth_state(thread_t th_act, void *tirp) { vm_offset_t header; int hoffset, i ; mythread_state_flavor_t *flavors; struct thread_command *tc; + tir_t *t = (tir_t *)tirp; + /* * Fill in thread command structure. */ @@ -116,95 +169,141 @@ collectth_state(thread_act_t th_act, tir_t *t) * the appropriate thread state struct for each * thread state flavor. */ - for (i = 0; i < mynum_flavors; i++) { + for (i = 0; i < t->flavor_count; i++) { *(mythread_state_flavor_t *)(header+hoffset) = flavors[i]; hoffset += sizeof(mythread_state_flavor_t); thread_getstatus(th_act, flavors[i].flavor, - (thread_state_t *)(header+hoffset), + (thread_state_t)(header+hoffset), &flavors[i].count); hoffset += flavors[i].count*sizeof(int); } t->hoffset = hoffset; } + + /* - * Create a core image on the file "core". + * coredump + * + * Description: Create a core image on the file "core" for the process + * indicated + * + * Parameters: core_proc Process to dump core [*] + * reserve_mb If non-zero, leave filesystem with + * at least this much free space. + * ignore_ulimit If set, ignore the process's core file ulimit. + * + * Returns: 0 Success + * EFAULT Failed + * + * IMPORTANT: This function can only be called on the current process, due + * to assumptions below; see variable declaration section for + * details. */ #define MAX_TSTATE_FLAVORS 10 int -coredump(p) - register struct proc *p; +coredump(proc_t core_proc, uint32_t reserve_mb, int ignore_ulimit) { - int error=0; - register struct pcred *pcred = p->p_cred; - register struct ucred *cred = pcred->pc_ucred; - struct nameidata nd; - struct vattr vattr; - vm_map_t map; +/* Begin assumptions that limit us to only the current process */ + vfs_context_t ctx = vfs_context_current(); + vm_map_t map = current_map(); + task_t task = current_task(); +/* End assumptions */ + kauth_cred_t cred = vfs_context_ucred(ctx); + int error = 0; + struct vnode_attr va; int thread_count, segment_count; int command_size, header_size, tstate_size; - int hoffset, foffset, vmoffset; + int hoffset; + off_t foffset; + mach_vm_offset_t vmoffset; vm_offset_t header; - struct machine_slot *ms; - struct mach_header *mh; - struct segment_command *sc; - struct thread_command *tc; - vm_size_t size; + mach_vm_size_t vmsize; vm_prot_t prot; vm_prot_t maxprot; vm_inherit_t inherit; - vm_offset_t offset; - int error1; - task_t task; - char core_name[MAXCOMLEN+6]; + int error1 = 0; + char stack_name[MAXCOMLEN+6]; + char *alloced_name = NULL; + char *name; mythread_state_flavor_t flavors[MAX_TSTATE_FLAVORS]; - vm_size_t nflavors,mapsize; + vm_size_t mapsize; int i; - int nesting_depth = 0; + uint32_t nesting_depth = 0; kern_return_t kret; struct vm_region_submap_info_64 vbr; - int vbrcount=0; + mach_msg_type_number_t vbrcount = 0; tir_t tir1; struct vnode * vp; + struct mach_header *mh = NULL; /* protected by is_64 */ + struct mach_header_64 *mh64 = NULL; /* protected by is_64 */ + int is_64 = 0; + size_t mach_header_sz = sizeof(struct mach_header); + size_t segment_command_sz = sizeof(struct segment_command); + + if (current_proc() != core_proc) { + panic("coredump() called against proc that is not current_proc: %p", core_proc); + } + if (do_coredump == 0 || /* Not dumping at all */ + ( (sugid_coredump == 0) && /* Not dumping SUID/SGID binaries */ + ( (kauth_cred_getsvuid(cred) != kauth_cred_getruid(cred)) || + (kauth_cred_getsvgid(cred) != kauth_cred_getrgid(cred))))) { - if (pcred->p_svuid != pcred->p_ruid || pcred->p_svgid != pcred->p_rgid) +#if CONFIG_AUDIT + audit_proc_coredump(core_proc, NULL, EFAULT); +#endif return (EFAULT); + } + + if (IS_64BIT_PROCESS(core_proc)) { + is_64 = 1; + mach_header_sz = sizeof(struct mach_header_64); + segment_command_sz = sizeof(struct segment_command_64); + } - task = current_task(); - map = current_map(); mapsize = get_vmmap_size(map); - if (mapsize >= p->p_rlimit[RLIMIT_CORE].rlim_cur) + if ((mapsize >= core_proc->p_rlimit[RLIMIT_CORE].rlim_cur) && (ignore_ulimit == 0)) return (EFAULT); - (void) task_suspend(task); + (void) task_suspend_internal(task); - /* - * Make sure all registers, etc. are in pcb so they get - * into core file. - */ -#if defined (__ppc__) - fpu_save(); -#endif - sprintf(core_name, "/cores/core.%d", p->p_pid); - NDINIT(&nd, LOOKUP, FOLLOW, UIO_SYSSPACE, core_name, p); - if(error = vn_open(&nd, O_CREAT | FWRITE, S_IRUSR )) - return (error); - vp = nd.ni_vp; - + MALLOC(alloced_name, char *, MAXPATHLEN, M_TEMP, M_NOWAIT | M_ZERO); + + /* create name according to sysctl'able format string */ + /* if name creation fails, fall back to historical behaviour... */ + if (alloced_name == NULL || + proc_core_name(core_proc->p_comm, kauth_cred_getuid(cred), + core_proc->p_pid, alloced_name, MAXPATHLEN)) { + snprintf(stack_name, sizeof(stack_name), + "/cores/core.%d", core_proc->p_pid); + name = stack_name; + } else + name = alloced_name; + + if ((error = vnode_open(name, (O_CREAT | FWRITE | O_NOFOLLOW), S_IRUSR, VNODE_LOOKUP_NOFOLLOW, &vp, ctx))) + goto out2; + + VATTR_INIT(&va); + VATTR_WANTED(&va, va_nlink); /* Don't dump to non-regular files or files with links. */ if (vp->v_type != VREG || - VOP_GETATTR(vp, &vattr, cred, p) || vattr.va_nlink != 1) { + vnode_getattr(vp, &va, ctx) || va.va_nlink != 1) { error = EFAULT; goto out; } - VATTR_NULL(&vattr); - vattr.va_size = 0; - VOP_LEASE(vp, p, cred, LEASE_WRITE); - VOP_SETATTR(vp, &vattr, cred, p); - p->p_acflag |= ACORE; + VATTR_INIT(&va); /* better to do it here than waste more stack in vnode_setsize */ + VATTR_SET(&va, va_data_size, 0); + vnode_setattr(vp, &va, ctx); + core_proc->p_acflag |= ACORE; + + if ((reserve_mb > 0) && + ((freespace_mb(vp) - (mapsize >> 20)) < reserve_mb)) { + error = ENOSPC; + goto out; + } /* * If the task is modified while dumping the file @@ -214,65 +313,78 @@ coredump(p) thread_count = get_task_numacts(task); segment_count = get_vmmap_entries(map); /* XXX */ - /* - * nflavors here is really the number of ints in flavors - * to meet the thread_getstatus() calling convention - */ -#if 0 - nflavors = sizeof(flavors)/sizeof(int); - if (thread_getstatus(current_thread(), THREAD_STATE_FLAVOR_LIST, - (thread_state_t)(flavors), - &nflavors) != KERN_SUCCESS) - panic("core flavor list"); - /* now convert to number of flavors */ - nflavors /= sizeof(mythread_state_flavor_t)/sizeof(int); -#else - nflavors = mynum_flavors; - bcopy(thread_flavor_array,flavors,sizeof(thread_flavor_array)); -#endif + tir1.flavor_count = sizeof(thread_flavor_array)/sizeof(mythread_state_flavor_t); + bcopy(thread_flavor_array, flavors,sizeof(thread_flavor_array)); tstate_size = 0; - for (i = 0; i < nflavors; i++) + for (i = 0; i < tir1.flavor_count; i++) tstate_size += sizeof(mythread_state_flavor_t) + (flavors[i].count * sizeof(int)); - - command_size = segment_count*sizeof(struct segment_command) + + command_size = segment_count * segment_command_sz + thread_count*sizeof(struct thread_command) + tstate_size*thread_count; - header_size = command_size + sizeof(struct mach_header); + header_size = command_size + mach_header_sz; - (void) kmem_alloc_wired(kernel_map, - (vm_offset_t *)&header, - (vm_size_t)header_size); + if (kmem_alloc(kernel_map, &header, (vm_size_t)header_size) != KERN_SUCCESS) { + error = ENOMEM; + goto out; + } /* * Set up Mach-O header. */ - mh = (struct mach_header *) header; - ms = &machine_slot[cpu_number()]; - mh->magic = MH_MAGIC; - mh->cputype = ms->cpu_type; - mh->cpusubtype = ms->cpu_subtype; - mh->filetype = MH_CORE; - mh->ncmds = segment_count + thread_count; - mh->sizeofcmds = command_size; - - hoffset = sizeof(struct mach_header); /* offset into header */ + if (is_64) { + mh64 = (struct mach_header_64 *)header; + mh64->magic = MH_MAGIC_64; + mh64->cputype = process_cpu_type(core_proc); + mh64->cpusubtype = process_cpu_subtype(core_proc); + mh64->filetype = MH_CORE; + mh64->ncmds = segment_count + thread_count; + mh64->sizeofcmds = command_size; + mh64->reserved = 0; /* 8 byte alignment */ + } else { + mh = (struct mach_header *)header; + mh->magic = MH_MAGIC; + mh->cputype = process_cpu_type(core_proc); + mh->cpusubtype = process_cpu_subtype(core_proc); + mh->filetype = MH_CORE; + mh->ncmds = segment_count + thread_count; + mh->sizeofcmds = command_size; + } + + hoffset = mach_header_sz; /* offset into header */ foffset = round_page(header_size); /* offset into file */ - vmoffset = VM_MIN_ADDRESS; /* offset into VM */ - /* We use to check for an error, here, now we try and get + vmoffset = MACH_VM_MIN_ADDRESS; /* offset into VM */ + + /* + * We use to check for an error, here, now we try and get * as much as we can */ - while (segment_count > 0){ + while (segment_count > 0) { + struct segment_command *sc; + struct segment_command_64 *sc64; + /* * Get region information for next region. */ while (1) { vbrcount = VM_REGION_SUBMAP_INFO_COUNT_64; - if((kret = vm_region_recurse_64(map, - &vmoffset, &size, &nesting_depth, - &vbr, &vbrcount)) != KERN_SUCCESS) { + if((kret = mach_vm_region_recurse(map, + &vmoffset, &vmsize, &nesting_depth, + (vm_region_recurse_info_t)&vbr, + &vbrcount)) != KERN_SUCCESS) { + break; + } + /* + * If we get a valid mapping back, but we're dumping + * a 32 bit process, and it's over the allowable + * address space of a 32 bit process, it's the same + * as if mach_vm_region_recurse() failed. + */ + if (!(is_64) && + (vmoffset + vmsize > VM_MAX_ADDRESS)) { + kret = KERN_INVALID_ADDRESS; break; } if(vbr.is_submap) { @@ -291,90 +403,101 @@ coredump(p) /* * Fill in segment command structure. */ - sc = (struct segment_command *) (header + hoffset); - sc->cmd = LC_SEGMENT; - sc->cmdsize = sizeof(struct segment_command); - /* segment name is zerod by kmem_alloc */ - sc->vmaddr = vmoffset; - sc->vmsize = size; - sc->fileoff = foffset; - sc->filesize = size; - sc->maxprot = maxprot; - sc->initprot = prot; - sc->nsects = 0; + if (is_64) { + sc64 = (struct segment_command_64 *)(header + hoffset); + sc64->cmd = LC_SEGMENT_64; + sc64->cmdsize = sizeof(struct segment_command_64); + /* segment name is zeroed by kmem_alloc */ + sc64->segname[0] = 0; + sc64->vmaddr = vmoffset; + sc64->vmsize = vmsize; + sc64->fileoff = foffset; + sc64->filesize = vmsize; + sc64->maxprot = maxprot; + sc64->initprot = prot; + sc64->nsects = 0; + } else { + sc = (struct segment_command *) (header + hoffset); + sc->cmd = LC_SEGMENT; + sc->cmdsize = sizeof(struct segment_command); + /* segment name is zeroed by kmem_alloc */ + sc->segname[0] = 0; + sc->vmaddr = CAST_DOWN_EXPLICIT(vm_offset_t,vmoffset); + sc->vmsize = CAST_DOWN_EXPLICIT(vm_size_t,vmsize); + sc->fileoff = CAST_DOWN_EXPLICIT(uint32_t,foffset); /* will never truncate */ + sc->filesize = CAST_DOWN_EXPLICIT(uint32_t,vmsize); /* will never truncate */ + sc->maxprot = maxprot; + sc->initprot = prot; + sc->nsects = 0; + } /* * Write segment out. Try as hard as possible to * get read access to the data. */ if ((prot & VM_PROT_READ) == 0) { - vm_protect(map, vmoffset, size, FALSE, - prot|VM_PROT_READ); + mach_vm_protect(map, vmoffset, vmsize, FALSE, + prot|VM_PROT_READ); } /* * Only actually perform write if we can read. * Note: if we can't read, then we end up with * a hole in the file. */ - if ((maxprot & VM_PROT_READ) == VM_PROT_READ) { - error = vn_rdwr(UIO_WRITE, vp, (caddr_t)vmoffset, size, foffset, - UIO_USERSPACE, IO_NODELOCKED|IO_UNIT, cred, (int *) 0, p); + if ((maxprot & VM_PROT_READ) == VM_PROT_READ + && vbr.user_tag != VM_MEMORY_IOKIT + && coredumpok(map,vmoffset)) { + + error = vn_rdwr_64(UIO_WRITE, vp, vmoffset, vmsize, foffset, + (IS_64BIT_PROCESS(core_proc) ? UIO_USERSPACE64 : UIO_USERSPACE32), + IO_NOCACHE|IO_NODELOCKED|IO_UNIT, cred, (int64_t *) 0, core_proc); + } - hoffset += sizeof(struct segment_command); - foffset += size; - vmoffset += size; + hoffset += segment_command_sz; + foffset += vmsize; + vmoffset += vmsize; segment_count--; } -#if 0 /* [ */ - task_lock(task); - thread = (thread_t) queue_first(&task->thread_list); - while (thread_count > 0) { - /* - * Fill in thread command structure. - */ - tc = (struct thread_command *) (header + hoffset); - tc->cmd = LC_THREAD; - tc->cmdsize = sizeof(struct thread_command) - + tstate_size; - hoffset += sizeof(struct thread_command); - /* - * Follow with a struct thread_state_flavor and - * the appropriate thread state struct for each - * thread state flavor. - */ - for (i = 0; i < nflavors; i++) { - *(mythread_state_flavor_t *)(header+hoffset) = - flavors[i]; - hoffset += sizeof(mythread_state_flavor_t); - thread_getstatus(thread, flavors[i].flavor, - (thread_state_t *)(header+hoffset), - &flavors[i].count); - hoffset += flavors[i].count*sizeof(int); - } - thread = (thread_t) queue_next(&thread->thread_list); - thread_count--; + /* + * If there are remaining segments which have not been written + * out because break in the loop above, then they were not counted + * because they exceed the real address space of the executable + * type: remove them from the header's count. This is OK, since + * we are allowed to have a sparse area following the segments. + */ + if (is_64) { + mh64->ncmds -= segment_count; + mh64->sizeofcmds -= segment_count * segment_command_sz; + } else { + mh->ncmds -= segment_count; + mh->sizeofcmds -= segment_count * segment_command_sz; } - task_unlock(task); -#else /* /* 0 ][ */ + tir1.header = header; tir1.hoffset = hoffset; tir1.flavors = flavors; tir1.tstate_size = tstate_size; task_act_iterate_wth_args(task, collectth_state,&tir1); -#endif /* 0 ] */ /* * Write out the Mach header at the beginning of the - * file. + * file. OK to use a 32 bit write for this. */ error = vn_rdwr(UIO_WRITE, vp, (caddr_t)header, header_size, (off_t)0, - UIO_SYSSPACE, IO_NODELOCKED|IO_UNIT, cred, (int *) 0, p); + UIO_SYSSPACE, IO_NOCACHE|IO_NODELOCKED|IO_UNIT, cred, (int *) 0, core_proc); kmem_free(kernel_map, header, header_size); out: - VOP_UNLOCK(vp, 0, p); - error1 = vn_close(vp, FWRITE, cred, p); + error1 = vnode_close(vp, FWRITE, ctx); +out2: +#if CONFIG_AUDIT + audit_proc_coredump(core_proc, name, error); +#endif + if (alloced_name != NULL) + FREE(alloced_name, M_TEMP); if (error == 0) error = error1; + + return (error); }