X-Git-Url: https://git.saurik.com/apple/xnu.git/blobdiff_plain/e5568f75972dfc723778653c11cb6b4dc825716a..a991bd8d3e7fe02dbca0644054bab73c5b75324a:/bsd/kern/kern_core.c diff --git a/bsd/kern/kern_core.c b/bsd/kern/kern_core.c index fe156fcbf..a4a3ee6cf 100644 --- a/bsd/kern/kern_core.c +++ b/bsd/kern/kern_core.c @@ -1,23 +1,29 @@ /* - * Copyright (c) 2000-2002 Apple Computer, Inc. All rights reserved. + * Copyright (c) 2000-2006 Apple Computer, Inc. All rights reserved. * - * @APPLE_LICENSE_HEADER_START@ - * - * The contents of this file constitute Original Code as defined in and - * are subject to the Apple Public Source License Version 1.1 (the - * "License"). You may not use this file except in compliance with the - * License. Please obtain a copy of the License at - * http://www.apple.com/publicsource and read it before using this file. - * - * This Original Code and all software distributed under the License are - * distributed on an "AS IS" basis, WITHOUT WARRANTY OF ANY KIND, EITHER + * @APPLE_OSREFERENCE_LICENSE_HEADER_START@ + * + * This file contains Original Code and/or Modifications of Original Code + * as defined in and that are subject to the Apple Public Source License + * Version 2.0 (the 'License'). You may not use this file except in + * compliance with the License. The rights granted to you under the License + * may not be used to create, or enable the creation or redistribution of, + * unlawful or unlicensed copies of an Apple operating system, or to + * circumvent, violate, or enable the circumvention or violation of, any + * terms of an Apple operating system software license agreement. + * + * Please obtain a copy of the License at + * http://www.opensource.apple.com/apsl/ and read it before using this file. + * + * The Original Code and all software distributed under the License are + * distributed on an 'AS IS' basis, WITHOUT WARRANTY OF ANY KIND, EITHER * EXPRESS OR IMPLIED, AND APPLE HEREBY DISCLAIMS ALL SUCH WARRANTIES, * INCLUDING WITHOUT LIMITATION, ANY WARRANTIES OF MERCHANTABILITY, - * FITNESS FOR A PARTICULAR PURPOSE OR NON-INFRINGEMENT. Please see the - * License for the specific language governing rights and limitations - * under the License. - * - * @APPLE_LICENSE_HEADER_END@ + * FITNESS FOR A PARTICULAR PURPOSE, QUIET ENJOYMENT OR NON-INFRINGEMENT. + * Please see the License for the specific language governing rights and + * limitations under the License. + * + * @APPLE_OSREFERENCE_LICENSE_HEADER_END@ */ /* Copyright (c) 1991 NeXT Computer, Inc. All rights reserved. * @@ -26,22 +32,23 @@ * This file contains machine independent code for performing core dumps. * */ +#if CONFIG_COREDUMP #include #include - +#include #include #include #include #include #include -#include -#include +#include +#include +#include #include #include -#include #include -#include +#include #include #include #include @@ -51,178 +58,290 @@ #include #include +#include /* last */ +#include /* current_map() */ +#include /* mach_vm_region_recurse() */ +#include /* task_suspend() */ +#include /* get_task_numacts() */ + +#include + +#if CONFIG_MACF +#include +#endif /* CONFIG_MACF */ typedef struct { - int flavor; /* the number for this flavor */ - int count; /* count of ints in this flavor */ + int flavor; /* the number for this flavor */ + mach_msg_type_number_t count; /* count of ints in this flavor */ } mythread_state_flavor_t; -#if defined (__ppc__) - -mythread_state_flavor_t thread_flavor_array[]={ - {PPC_THREAD_STATE , PPC_THREAD_STATE_COUNT}, - {PPC_FLOAT_STATE, PPC_FLOAT_STATE_COUNT}, - {PPC_EXCEPTION_STATE, PPC_EXCEPTION_STATE_COUNT}, - {PPC_VECTOR_STATE, PPC_VECTOR_STATE_COUNT} - }; -int mynum_flavors=4; -#elif defined (__i386__) -mythread_state_flavor_t thread_flavor_array [] = { - {i386_THREAD_STATE, i386_THREAD_STATE_COUNT}, - {i386_THREAD_FPSTATE, i386_THREAD_FPSTATE_COUNT}, - {i386_THREAD_EXCEPTSTATE, i386_THREAD_EXCEPTSTATE_COUNT}, - {i386_THREAD_CTHREADSTATE, i386_THREAD_CTHREADSTATE_COUNT}, - {i386_NEW_THREAD_STATE, i386_NEW_THREAD_STATE_COUNT}, - {i386_FLOAT_STATE, i386_FLOAT_STATE_COUNT}, - {i386_ISA_PORT_MAP_STATE, i386_ISA_PORT_MAP_STATE_COUNT}, - {i386_V86_ASSIST_STATE, i386_V86_ASSIST_STATE_COUNT}, - {THREAD_SYSCALL_STATE, i386_THREAD_SYSCALL_STATE_COUNT} - }; -int mynum_flavors=9; - +#if defined (__i386__) || defined (__x86_64__) +mythread_state_flavor_t thread_flavor_array[] = { + {x86_THREAD_STATE, x86_THREAD_STATE_COUNT}, + {x86_FLOAT_STATE, x86_FLOAT_STATE_COUNT}, + {x86_EXCEPTION_STATE, x86_EXCEPTION_STATE_COUNT}, +}; +int mynum_flavors = 3; +#elif defined (__arm__) +mythread_state_flavor_t thread_flavor_array[] = { + {ARM_THREAD_STATE, ARM_THREAD_STATE_COUNT}, + {ARM_VFP_STATE, ARM_VFP_STATE_COUNT}, + {ARM_EXCEPTION_STATE, ARM_EXCEPTION_STATE_COUNT} +}; +int mynum_flavors = 3; + +#elif defined (__arm64__) +mythread_state_flavor_t thread_flavor_array[] = { + {ARM_THREAD_STATE64, ARM_THREAD_STATE64_COUNT}, + /* ARM64_TODO: VFP */ + {ARM_EXCEPTION_STATE64, ARM_EXCEPTION_STATE64_COUNT} +}; +int mynum_flavors = 2; #else #error architecture not supported #endif typedef struct { - vm_offset_t header; - int hoffset; + vm_offset_t header; + size_t hoffset; mythread_state_flavor_t *flavors; - int tstate_size; + size_t tstate_size; + size_t flavor_count; } tir_t; -/* XXX should be static */ -void collectth_state(thread_act_t th_act, tir_t *t); +extern int freespace_mb(vnode_t vp); /* XXX not in a Mach header anywhere */ -kern_return_t thread_getstatus(register thread_act_t act, int flavor, - thread_state_t tstate, mach_msg_type_number_t *count); +kern_return_t thread_getstatus(thread_t act, int flavor, + thread_state_t tstate, mach_msg_type_number_t *count); +void task_act_iterate_wth_args(task_t, void (*)(thread_t, void *), void *); + +#ifdef SECURE_KERNEL +__XNU_PRIVATE_EXTERN int do_coredump = 0; /* default: don't dump cores */ +#else +__XNU_PRIVATE_EXTERN int do_coredump = 1; /* default: dump cores */ +#endif +__XNU_PRIVATE_EXTERN int sugid_coredump = 0; /* default: but not SGUID binaries */ -__private_extern__ do_coredump = 1; /* default: dump cores */ -__private_extern__ sugid_coredump = 0; /* deafult: but not on SGUID binaries */ +/* cpu_type returns only the most generic indication of the current CPU. */ +/* in a core we want to know the kind of process. */ -void -collectth_state(thread_act_t th_act, tir_t *t) +static cpu_type_t +process_cpu_type(proc_t core_proc) { - vm_offset_t header; - int hoffset, i ; - mythread_state_flavor_t *flavors; - struct thread_command *tc; - /* - * Fill in thread command structure. - */ - header = t->header; - hoffset = t->hoffset; - flavors = t->flavors; - - tc = (struct thread_command *) (header + hoffset); - tc->cmd = LC_THREAD; - tc->cmdsize = sizeof(struct thread_command) - + t->tstate_size; - hoffset += sizeof(struct thread_command); - /* - * Follow with a struct thread_state_flavor and - * the appropriate thread state struct for each - * thread state flavor. - */ - for (i = 0; i < mynum_flavors; i++) { - *(mythread_state_flavor_t *)(header+hoffset) = - flavors[i]; - hoffset += sizeof(mythread_state_flavor_t); - thread_getstatus(th_act, flavors[i].flavor, - (thread_state_t)(header+hoffset), - &flavors[i].count); - hoffset += flavors[i].count*sizeof(int); - } + cpu_type_t what_we_think; +#if defined (__i386__) || defined (__x86_64__) + if (IS_64BIT_PROCESS(core_proc)) { + what_we_think = CPU_TYPE_X86_64; + } else { + what_we_think = CPU_TYPE_I386; + } +#elif defined (__arm__) || defined(__arm64__) + if (IS_64BIT_PROCESS(core_proc)) { + what_we_think = CPU_TYPE_ARM64; + } else { + what_we_think = CPU_TYPE_ARM; + } +#endif - t->hoffset = hoffset; + return what_we_think; } -extern boolean_t coredumpok(vm_map_t map, vm_offset_t va); /* temp fix */ -extern task_t current_task(void); /* XXX */ +static cpu_type_t +process_cpu_subtype(proc_t core_proc) +{ + cpu_type_t what_we_think; +#if defined (__i386__) || defined (__x86_64__) + if (IS_64BIT_PROCESS(core_proc)) { + what_we_think = CPU_SUBTYPE_X86_64_ALL; + } else { + what_we_think = CPU_SUBTYPE_I386_ALL; + } +#elif defined (__arm__) || defined(__arm64__) + if (IS_64BIT_PROCESS(core_proc)) { + what_we_think = CPU_SUBTYPE_ARM64_ALL; + } else { + what_we_think = CPU_SUBTYPE_ARM_ALL; + } +#endif + return what_we_think; +} + +static void +collectth_state(thread_t th_act, void *tirp) +{ + vm_offset_t header; + size_t hoffset, i; + mythread_state_flavor_t *flavors; + struct thread_command *tc; + tir_t *t = (tir_t *)tirp; + + /* + * Fill in thread command structure. + */ + header = t->header; + hoffset = t->hoffset; + flavors = t->flavors; + + tc = (struct thread_command *) (header + hoffset); + tc->cmd = LC_THREAD; + tc->cmdsize = (uint32_t)(sizeof(struct thread_command) + + t->tstate_size); + hoffset += sizeof(struct thread_command); + /* + * Follow with a struct thread_state_flavor and + * the appropriate thread state struct for each + * thread state flavor. + */ + for (i = 0; i < t->flavor_count; i++) { + *(mythread_state_flavor_t *)(header + hoffset) = + flavors[i]; + hoffset += sizeof(mythread_state_flavor_t); + thread_getstatus(th_act, flavors[i].flavor, + (thread_state_t)(header + hoffset), + &flavors[i].count); + hoffset += flavors[i].count * sizeof(int); + } + + t->hoffset = hoffset; +} /* - * Create a core image on the file "core". + * coredump + * + * Description: Create a core image on the file "core" for the process + * indicated + * + * Parameters: core_proc Process to dump core [*] + * reserve_mb If non-zero, leave filesystem with + * at least this much free space. + * coredump_flags Extra options (ignore rlimit, run fsync) + * + * Returns: 0 Success + * !0 Failure errno + * + * IMPORTANT: This function can only be called on the current process, due + * to assumptions below; see variable declaration section for + * details. */ -#define MAX_TSTATE_FLAVORS 10 +#define MAX_TSTATE_FLAVORS 10 int -coredump(struct proc *p) +coredump(proc_t core_proc, uint32_t reserve_mb, int coredump_flags) { - int error=0; - register struct pcred *pcred = p->p_cred; - register struct ucred *cred = pcred->pc_ucred; - struct nameidata nd; - struct vattr vattr; - vm_map_t map; - int thread_count, segment_count; - int command_size, header_size, tstate_size; - int hoffset, foffset, vmoffset; - vm_offset_t header; - struct machine_slot *ms; - struct mach_header *mh; - struct segment_command *sc; - vm_size_t size; - vm_prot_t prot; - vm_prot_t maxprot; - vm_inherit_t inherit; - int error1; - task_t task; - char core_name[MAXCOMLEN+6]; - char *name; +/* Begin assumptions that limit us to only the current process */ + vfs_context_t ctx = vfs_context_current(); + vm_map_t map = current_map(); + task_t task = current_task(); +/* End assumptions */ + kauth_cred_t cred = vfs_context_ucred(ctx); + int error = 0; + struct vnode_attr va; + size_t thread_count, segment_count; + size_t command_size, header_size, tstate_size; + size_t hoffset; + off_t foffset; + mach_vm_offset_t vmoffset; + vm_offset_t header; + mach_vm_size_t vmsize; + vm_prot_t prot; + vm_prot_t maxprot; + vm_inherit_t inherit; + int error1 = 0; + char stack_name[MAXCOMLEN + 6]; + char *alloced_name = NULL; + char *name = NULL; mythread_state_flavor_t flavors[MAX_TSTATE_FLAVORS]; - vm_size_t mapsize; - int i; - int nesting_depth = 0; - kern_return_t kret; + vm_size_t mapsize; + size_t i; + uint32_t nesting_depth = 0; + kern_return_t kret; struct vm_region_submap_info_64 vbr; - int vbrcount=0; + mach_msg_type_number_t vbrcount = 0; tir_t tir1; struct vnode * vp; + struct mach_header *mh = NULL; /* protected by is_64 */ + struct mach_header_64 *mh64 = NULL; /* protected by is_64 */ + int is_64 = 0; + size_t mach_header_sz = sizeof(struct mach_header); + size_t segment_command_sz = sizeof(struct segment_command); + + if (current_proc() != core_proc) { + panic("coredump() called against proc that is not current_proc: %p", core_proc); + } + + if (do_coredump == 0 || /* Not dumping at all */ + ((sugid_coredump == 0) && /* Not dumping SUID/SGID binaries */ + ((kauth_cred_getsvuid(cred) != kauth_cred_getruid(cred)) || + (kauth_cred_getsvgid(cred) != kauth_cred_getrgid(cred))))) { + error = EFAULT; + goto out2; + } + +#if CONFIG_MACF + error = mac_proc_check_dump_core(core_proc); + if (error != 0) { + goto out2; + } +#endif - if (do_coredump == 0 || /* Not dumping at all */ - ( (sugid_coredump == 0) && /* Not dumping SUID/SGID binaries */ - ( (pcred->p_svuid != pcred->p_ruid) || - (pcred->p_svgid != pcred->p_rgid)))) { - - return (EFAULT); + if (IS_64BIT_PROCESS(core_proc)) { + is_64 = 1; + mach_header_sz = sizeof(struct mach_header_64); + segment_command_sz = sizeof(struct segment_command_64); } - task = current_task(); - map = current_map(); mapsize = get_vmmap_size(map); - if (mapsize >= p->p_rlimit[RLIMIT_CORE].rlim_cur) - return (EFAULT); - (void) task_suspend(task); + if (((coredump_flags & COREDUMP_IGNORE_ULIMIT) == 0) && + (mapsize >= proc_limitgetcur(core_proc, RLIMIT_CORE, FALSE))) { + error = EFAULT; + goto out2; + } - /* create name according to sysctl'able format string */ - name = proc_core_name(p->p_comm, p->p_ucred->cr_uid, p->p_pid); + (void) task_suspend_internal(task); + MALLOC(alloced_name, char *, MAXPATHLEN, M_TEMP, M_NOWAIT | M_ZERO); + + /* create name according to sysctl'able format string */ /* if name creation fails, fall back to historical behaviour... */ - if (name == NULL) { - sprintf(core_name, "/cores/core.%d", p->p_pid); - name = core_name; + if (alloced_name == NULL || + proc_core_name(core_proc->p_comm, kauth_cred_getuid(cred), + core_proc->p_pid, alloced_name, MAXPATHLEN)) { + snprintf(stack_name, sizeof(stack_name), + "/cores/core.%d", core_proc->p_pid); + name = stack_name; + } else { + name = alloced_name; + } + + if ((error = vnode_open(name, (O_CREAT | FWRITE | O_NOFOLLOW), S_IRUSR, VNODE_LOOKUP_NOFOLLOW, &vp, ctx))) { + goto out2; } - NDINIT(&nd, LOOKUP, NOFOLLOW, UIO_SYSSPACE, name, p); - if((error = vn_open(&nd, O_CREAT | FWRITE | O_NOFOLLOW, S_IRUSR )) != 0) - return (error); - vp = nd.ni_vp; - + VATTR_INIT(&va); + VATTR_WANTED(&va, va_nlink); /* Don't dump to non-regular files or files with links. */ if (vp->v_type != VREG || - VOP_GETATTR(vp, &vattr, cred, p) || vattr.va_nlink != 1) { + vnode_getattr(vp, &va, ctx) || va.va_nlink != 1) { error = EFAULT; goto out; } - VATTR_NULL(&vattr); - vattr.va_size = 0; - VOP_LEASE(vp, p, cred, LEASE_WRITE); - VOP_SETATTR(vp, &vattr, cred, p); - p->p_acflag |= ACORE; + VATTR_INIT(&va); /* better to do it here than waste more stack in vnode_setsize */ + VATTR_SET(&va, va_data_size, 0); + if (core_proc == initproc) { + VATTR_SET(&va, va_dataprotect_class, PROTECTION_CLASS_D); + } + vnode_setattr(vp, &va, ctx); + core_proc->p_acflag |= ACORE; + + if ((reserve_mb > 0) && + ((freespace_mb(vp) - (mapsize >> 20)) < reserve_mb)) { + error = ENOSPC; + goto out; + } /* * If the task is modified while dumping the file @@ -231,63 +350,115 @@ coredump(struct proc *p) */ thread_count = get_task_numacts(task); - segment_count = get_vmmap_entries(map); /* XXX */ - bcopy(thread_flavor_array,flavors,sizeof(thread_flavor_array)); + segment_count = get_vmmap_entries(map); /* XXX */ + tir1.flavor_count = sizeof(thread_flavor_array) / sizeof(mythread_state_flavor_t); + bcopy(thread_flavor_array, flavors, sizeof(thread_flavor_array)); tstate_size = 0; - for (i = 0; i < mynum_flavors; i++) + for (i = 0; i < tir1.flavor_count; i++) { tstate_size += sizeof(mythread_state_flavor_t) + - (flavors[i].count * sizeof(int)); + (flavors[i].count * sizeof(int)); + } - command_size = segment_count*sizeof(struct segment_command) + - thread_count*sizeof(struct thread_command) + - tstate_size*thread_count; + { + size_t lhs; + size_t rhs; - header_size = command_size + sizeof(struct mach_header); + /* lhs = segment_count * segment_command_sz */ + if (os_mul_overflow(segment_count, segment_command_sz, &lhs)) { + error = ENOMEM; + goto out; + } + + /* rhs = (tstate_size + sizeof(struct thread_command)) * thread_count */ + if (os_add_and_mul_overflow(tstate_size, sizeof(struct thread_command), thread_count, &rhs)) { + error = ENOMEM; + goto out; + } + + /* command_size = lhs + rhs */ + if (os_add_overflow(lhs, rhs, &command_size)) { + error = ENOMEM; + goto out; + } + } - (void) kmem_alloc_wired(kernel_map, - (vm_offset_t *)&header, - (vm_size_t)header_size); + if (os_add_overflow(command_size, mach_header_sz, &header_size)) { + error = ENOMEM; + goto out; + } + + if (kmem_alloc(kernel_map, &header, (vm_size_t)header_size, VM_KERN_MEMORY_DIAG) != KERN_SUCCESS) { + error = ENOMEM; + goto out; + } /* * Set up Mach-O header. */ - mh = (struct mach_header *) header; - ms = &machine_slot[cpu_number()]; - mh->magic = MH_MAGIC; - mh->cputype = ms->cpu_type; - mh->cpusubtype = ms->cpu_subtype; - mh->filetype = MH_CORE; - mh->ncmds = segment_count + thread_count; - mh->sizeofcmds = command_size; - - hoffset = sizeof(struct mach_header); /* offset into header */ - foffset = round_page_32(header_size); /* offset into file */ - vmoffset = VM_MIN_ADDRESS; /* offset into VM */ + if (is_64) { + mh64 = (struct mach_header_64 *)header; + mh64->magic = MH_MAGIC_64; + mh64->cputype = process_cpu_type(core_proc); + mh64->cpusubtype = process_cpu_subtype(core_proc); + mh64->filetype = MH_CORE; + mh64->ncmds = (uint32_t)(segment_count + thread_count); + mh64->sizeofcmds = (uint32_t)command_size; + mh64->reserved = 0; /* 8 byte alignment */ + } else { + mh = (struct mach_header *)header; + mh->magic = MH_MAGIC; + mh->cputype = process_cpu_type(core_proc); + mh->cpusubtype = process_cpu_subtype(core_proc); + mh->filetype = MH_CORE; + mh->ncmds = (uint32_t)(segment_count + thread_count); + mh->sizeofcmds = (uint32_t)command_size; + } + + hoffset = mach_header_sz; /* offset into header */ + foffset = round_page(header_size); /* offset into file */ + vmoffset = MACH_VM_MIN_ADDRESS; /* offset into VM */ + /* - * We use to check for an error, here, now we try and get + * We use to check for an error, here, now we try and get * as much as we can */ - while (segment_count > 0){ + while (segment_count > 0) { + struct segment_command *sc; + struct segment_command_64 *sc64; + /* * Get region information for next region. */ - + while (1) { vbrcount = VM_REGION_SUBMAP_INFO_COUNT_64; - if((kret = vm_region_recurse_64(map, - &vmoffset, &size, &nesting_depth, - &vbr, &vbrcount)) != KERN_SUCCESS) { + if ((kret = mach_vm_region_recurse(map, + &vmoffset, &vmsize, &nesting_depth, + (vm_region_recurse_info_t)&vbr, + &vbrcount)) != KERN_SUCCESS) { + break; + } + /* + * If we get a valid mapping back, but we're dumping + * a 32 bit process, and it's over the allowable + * address space of a 32 bit process, it's the same + * as if mach_vm_region_recurse() failed. + */ + if (!(is_64) && + (vmoffset + vmsize > VM_MAX_ADDRESS)) { + kret = KERN_INVALID_ADDRESS; break; } - if(vbr.is_submap) { + if (vbr.is_submap) { nesting_depth++; continue; } else { break; } } - if(kret != KERN_SUCCESS) + if (kret != KERN_SUCCESS) { break; + } prot = vbr.protection; maxprot = vbr.max_protection; @@ -295,26 +466,43 @@ coredump(struct proc *p) /* * Fill in segment command structure. */ - sc = (struct segment_command *) (header + hoffset); - sc->cmd = LC_SEGMENT; - sc->cmdsize = sizeof(struct segment_command); - /* segment name is zerod by kmem_alloc */ - sc->segname[0] = 0; - sc->vmaddr = vmoffset; - sc->vmsize = size; - sc->fileoff = foffset; - sc->filesize = size; - sc->maxprot = maxprot; - sc->initprot = prot; - sc->nsects = 0; + if (is_64) { + sc64 = (struct segment_command_64 *)(header + hoffset); + sc64->cmd = LC_SEGMENT_64; + sc64->cmdsize = sizeof(struct segment_command_64); + /* segment name is zeroed by kmem_alloc */ + sc64->segname[0] = 0; + sc64->vmaddr = vmoffset; + sc64->vmsize = vmsize; + sc64->fileoff = foffset; + sc64->filesize = vmsize; + sc64->maxprot = maxprot; + sc64->initprot = prot; + sc64->nsects = 0; + sc64->flags = 0; + } else { + sc = (struct segment_command *) (header + hoffset); + sc->cmd = LC_SEGMENT; + sc->cmdsize = sizeof(struct segment_command); + /* segment name is zeroed by kmem_alloc */ + sc->segname[0] = 0; + sc->vmaddr = CAST_DOWN_EXPLICIT(uint32_t, vmoffset); + sc->vmsize = CAST_DOWN_EXPLICIT(uint32_t, vmsize); + sc->fileoff = CAST_DOWN_EXPLICIT(uint32_t, foffset); /* will never truncate */ + sc->filesize = CAST_DOWN_EXPLICIT(uint32_t, vmsize); /* will never truncate */ + sc->maxprot = maxprot; + sc->initprot = prot; + sc->nsects = 0; + sc->flags = 0; + } /* * Write segment out. Try as hard as possible to * get read access to the data. */ if ((prot & VM_PROT_READ) == 0) { - vm_protect(map, vmoffset, size, FALSE, - prot|VM_PROT_READ); + mach_vm_protect(map, vmoffset, vmsize, FALSE, + prot | VM_PROT_READ); } /* * Only actually perform write if we can read. @@ -322,35 +510,71 @@ coredump(struct proc *p) * a hole in the file. */ if ((maxprot & VM_PROT_READ) == VM_PROT_READ - && vbr.user_tag != VM_MEMORY_IOKIT - && coredumpok(map,vmoffset)) { - error = vn_rdwr(UIO_WRITE, vp, (caddr_t)vmoffset, size, foffset, - UIO_USERSPACE, IO_NODELOCKED|IO_UNIT, cred, (int *) 0, p); + && vbr.user_tag != VM_MEMORY_IOKIT + && coredumpok(map, vmoffset)) { + error = vn_rdwr_64(UIO_WRITE, vp, vmoffset, vmsize, foffset, + (IS_64BIT_PROCESS(core_proc) ? UIO_USERSPACE64 : UIO_USERSPACE32), + IO_NOCACHE | IO_NODELOCKED | IO_UNIT, cred, (int64_t *) 0, core_proc); } - hoffset += sizeof(struct segment_command); - foffset += size; - vmoffset += size; + hoffset += segment_command_sz; + foffset += vmsize; + vmoffset += vmsize; segment_count--; } + /* + * If there are remaining segments which have not been written + * out because break in the loop above, then they were not counted + * because they exceed the real address space of the executable + * type: remove them from the header's count. This is OK, since + * we are allowed to have a sparse area following the segments. + */ + if (is_64) { + mh64->ncmds -= segment_count; + mh64->sizeofcmds -= segment_count * segment_command_sz; + } else { + mh->ncmds -= segment_count; + mh->sizeofcmds -= segment_count * segment_command_sz; + } + tir1.header = header; tir1.hoffset = hoffset; tir1.flavors = flavors; tir1.tstate_size = tstate_size; - task_act_iterate_wth_args(task, collectth_state,&tir1); + task_act_iterate_wth_args(task, collectth_state, &tir1); /* * Write out the Mach header at the beginning of the - * file. + * file. OK to use a 32 bit write for this. */ - error = vn_rdwr(UIO_WRITE, vp, (caddr_t)header, header_size, (off_t)0, - UIO_SYSSPACE, IO_NODELOCKED|IO_UNIT, cred, (int *) 0, p); + error = vn_rdwr(UIO_WRITE, vp, (caddr_t)header, (int)MIN(header_size, INT_MAX), (off_t)0, + UIO_SYSSPACE, IO_NOCACHE | IO_NODELOCKED | IO_UNIT, cred, (int *) 0, core_proc); kmem_free(kernel_map, header, header_size); + + if ((coredump_flags & COREDUMP_FULLFSYNC) && error == 0) { + error = VNOP_IOCTL(vp, F_FULLFSYNC, (caddr_t)NULL, 0, ctx); + } out: - VOP_UNLOCK(vp, 0, p); - error1 = vn_close(vp, FWRITE, cred, p); - if (error == 0) + error1 = vnode_close(vp, FWRITE, ctx); +out2: +#if CONFIG_AUDIT + audit_proc_coredump(core_proc, name, error); +#endif + if (alloced_name != NULL) { + FREE(alloced_name, M_TEMP); + } + if (error == 0) { error = error1; - return (error); + } + + return error; } + +#else /* CONFIG_COREDUMP */ + +/* When core dumps aren't needed, no need to compile this file at all */ + +#error assertion failed: this section is not compiled + +#endif /* CONFIG_COREDUMP */