/*
- * Copyright (c) 2000-2002 Apple Computer, Inc. All rights reserved.
+ * Copyright (c) 2000-2006 Apple Computer, Inc. All rights reserved.
+ *
+ * @APPLE_OSREFERENCE_LICENSE_HEADER_START@
*
- * @APPLE_LICENSE_HEADER_START@
- *
- * Copyright (c) 1999-2003 Apple Computer, Inc. All Rights Reserved.
- *
* This file contains Original Code and/or Modifications of Original Code
* as defined in and that are subject to the Apple Public Source License
* Version 2.0 (the 'License'). You may not use this file except in
- * compliance with the License. Please obtain a copy of the License at
- * http://www.opensource.apple.com/apsl/ and read it before using this
- * file.
- *
+ * compliance with the License. The rights granted to you under the License
+ * may not be used to create, or enable the creation or redistribution of,
+ * unlawful or unlicensed copies of an Apple operating system, or to
+ * circumvent, violate, or enable the circumvention or violation of, any
+ * terms of an Apple operating system software license agreement.
+ *
+ * Please obtain a copy of the License at
+ * http://www.opensource.apple.com/apsl/ and read it before using this file.
+ *
* The Original Code and all software distributed under the License are
* distributed on an 'AS IS' basis, WITHOUT WARRANTY OF ANY KIND, EITHER
* EXPRESS OR IMPLIED, AND APPLE HEREBY DISCLAIMS ALL SUCH WARRANTIES,
* FITNESS FOR A PARTICULAR PURPOSE, QUIET ENJOYMENT OR NON-INFRINGEMENT.
* Please see the License for the specific language governing rights and
* limitations under the License.
- *
- * @APPLE_LICENSE_HEADER_END@
+ *
+ * @APPLE_OSREFERENCE_LICENSE_HEADER_END@
*/
/* Copyright (c) 1991 NeXT Computer, Inc. All rights reserved.
*
* This file contains machine independent code for performing core dumps.
*
*/
+#if CONFIG_COREDUMP
#include <mach/vm_param.h>
#include <mach/thread_status.h>
-
+#include <sys/content_protection.h>
#include <sys/param.h>
#include <sys/systm.h>
#include <sys/signalvar.h>
#include <sys/resourcevar.h>
#include <sys/namei.h>
-#include <sys/vnode.h>
-#include <sys/proc.h>
+#include <sys/vnode_internal.h>
+#include <sys/proc_internal.h>
+#include <sys/kauth.h>
#include <sys/timeb.h>
#include <sys/times.h>
-#include <sys/buf.h>
#include <sys/acct.h>
-#include <sys/file.h>
+#include <sys/file_internal.h>
#include <sys/uio.h>
#include <sys/kernel.h>
#include <sys/stat.h>
#include <mach/vm_statistics.h>
#include <vm/vm_kern.h>
+#include <vm/vm_protos.h> /* last */
+#include <vm/vm_map.h> /* current_map() */
+#include <mach/mach_vm.h> /* mach_vm_region_recurse() */
+#include <mach/task.h> /* task_suspend() */
+#include <kern/task.h> /* get_task_numacts() */
+
+#include <security/audit/audit.h>
+
+#if CONFIG_MACF
+#include <security/mac_framework.h>
+#endif /* CONFIG_MACF */
typedef struct {
- int flavor; /* the number for this flavor */
- int count; /* count of ints in this flavor */
+ int flavor; /* the number for this flavor */
+ mach_msg_type_number_t count; /* count of ints in this flavor */
} mythread_state_flavor_t;
-#if defined (__ppc__)
-
-mythread_state_flavor_t thread_flavor_array[]={
- {PPC_THREAD_STATE , PPC_THREAD_STATE_COUNT},
- {PPC_FLOAT_STATE, PPC_FLOAT_STATE_COUNT},
- {PPC_EXCEPTION_STATE, PPC_EXCEPTION_STATE_COUNT},
- {PPC_VECTOR_STATE, PPC_VECTOR_STATE_COUNT}
- };
-int mynum_flavors=4;
-#elif defined (__i386__)
-mythread_state_flavor_t thread_flavor_array [] = {
- {i386_THREAD_STATE, i386_THREAD_STATE_COUNT},
- {i386_THREAD_FPSTATE, i386_THREAD_FPSTATE_COUNT},
- {i386_THREAD_EXCEPTSTATE, i386_THREAD_EXCEPTSTATE_COUNT},
- {i386_THREAD_CTHREADSTATE, i386_THREAD_CTHREADSTATE_COUNT},
- {i386_NEW_THREAD_STATE, i386_NEW_THREAD_STATE_COUNT},
- {i386_FLOAT_STATE, i386_FLOAT_STATE_COUNT},
- {i386_ISA_PORT_MAP_STATE, i386_ISA_PORT_MAP_STATE_COUNT},
- {i386_V86_ASSIST_STATE, i386_V86_ASSIST_STATE_COUNT},
- {THREAD_SYSCALL_STATE, i386_THREAD_SYSCALL_STATE_COUNT}
- };
-int mynum_flavors=9;
+#if defined (__i386__) || defined (__x86_64__)
+mythread_state_flavor_t thread_flavor_array[] = {
+ {x86_THREAD_STATE, x86_THREAD_STATE_COUNT},
+ {x86_FLOAT_STATE, x86_FLOAT_STATE_COUNT},
+ {x86_EXCEPTION_STATE, x86_EXCEPTION_STATE_COUNT},
+};
+int mynum_flavors = 3;
+#elif defined (__arm__)
+mythread_state_flavor_t thread_flavor_array[] = {
+ {ARM_THREAD_STATE, ARM_THREAD_STATE_COUNT},
+ {ARM_VFP_STATE, ARM_VFP_STATE_COUNT},
+ {ARM_EXCEPTION_STATE, ARM_EXCEPTION_STATE_COUNT}
+};
+int mynum_flavors = 3;
+#elif defined (__arm64__)
+mythread_state_flavor_t thread_flavor_array[] = {
+ {ARM_THREAD_STATE64, ARM_THREAD_STATE64_COUNT},
+ /* ARM64_TODO: VFP */
+ {ARM_EXCEPTION_STATE64, ARM_EXCEPTION_STATE64_COUNT}
+};
+int mynum_flavors = 2;
#else
#error architecture not supported
#endif
typedef struct {
- vm_offset_t header;
- int hoffset;
+ vm_offset_t header;
+ size_t hoffset;
mythread_state_flavor_t *flavors;
- int tstate_size;
+ size_t tstate_size;
+ size_t flavor_count;
} tir_t;
-void
-collectth_state(thread_act_t th_act, tir_t *t)
+extern int freespace_mb(vnode_t vp);
+
+/* XXX not in a Mach header anywhere */
+kern_return_t thread_getstatus(thread_t act, int flavor,
+ thread_state_t tstate, mach_msg_type_number_t *count);
+void task_act_iterate_wth_args(task_t, void (*)(thread_t, void *), void *);
+
+#ifdef SECURE_KERNEL
+__XNU_PRIVATE_EXTERN int do_coredump = 0; /* default: don't dump cores */
+#else
+__XNU_PRIVATE_EXTERN int do_coredump = 1; /* default: dump cores */
+#endif
+__XNU_PRIVATE_EXTERN int sugid_coredump = 0; /* default: but not SGUID binaries */
+
+
+/* cpu_type returns only the most generic indication of the current CPU. */
+/* in a core we want to know the kind of process. */
+
+static cpu_type_t
+process_cpu_type(proc_t core_proc)
{
- vm_offset_t header;
- int hoffset, i ;
+ cpu_type_t what_we_think;
+#if defined (__i386__) || defined (__x86_64__)
+ if (IS_64BIT_PROCESS(core_proc)) {
+ what_we_think = CPU_TYPE_X86_64;
+ } else {
+ what_we_think = CPU_TYPE_I386;
+ }
+#elif defined (__arm__) || defined(__arm64__)
+ if (IS_64BIT_PROCESS(core_proc)) {
+ what_we_think = CPU_TYPE_ARM64;
+ } else {
+ what_we_think = CPU_TYPE_ARM;
+ }
+#endif
+
+ return what_we_think;
+}
+
+static cpu_type_t
+process_cpu_subtype(proc_t core_proc)
+{
+ cpu_type_t what_we_think;
+#if defined (__i386__) || defined (__x86_64__)
+ if (IS_64BIT_PROCESS(core_proc)) {
+ what_we_think = CPU_SUBTYPE_X86_64_ALL;
+ } else {
+ what_we_think = CPU_SUBTYPE_I386_ALL;
+ }
+#elif defined (__arm__) || defined(__arm64__)
+ if (IS_64BIT_PROCESS(core_proc)) {
+ what_we_think = CPU_SUBTYPE_ARM64_ALL;
+ } else {
+ what_we_think = CPU_SUBTYPE_ARM_ALL;
+ }
+#endif
+ return what_we_think;
+}
+
+static void
+collectth_state(thread_t th_act, void *tirp)
+{
+ vm_offset_t header;
+ size_t hoffset, i;
mythread_state_flavor_t *flavors;
- struct thread_command *tc;
- /*
- * Fill in thread command structure.
- */
- header = t->header;
- hoffset = t->hoffset;
- flavors = t->flavors;
-
- tc = (struct thread_command *) (header + hoffset);
- tc->cmd = LC_THREAD;
- tc->cmdsize = sizeof(struct thread_command)
- + t->tstate_size;
- hoffset += sizeof(struct thread_command);
- /*
- * Follow with a struct thread_state_flavor and
- * the appropriate thread state struct for each
- * thread state flavor.
- */
- for (i = 0; i < mynum_flavors; i++) {
- *(mythread_state_flavor_t *)(header+hoffset) =
- flavors[i];
- hoffset += sizeof(mythread_state_flavor_t);
- thread_getstatus(th_act, flavors[i].flavor,
- (thread_state_t *)(header+hoffset),
- &flavors[i].count);
- hoffset += flavors[i].count*sizeof(int);
- }
+ struct thread_command *tc;
+ tir_t *t = (tir_t *)tirp;
+
+ /*
+ * Fill in thread command structure.
+ */
+ header = t->header;
+ hoffset = t->hoffset;
+ flavors = t->flavors;
+
+ tc = (struct thread_command *) (header + hoffset);
+ tc->cmd = LC_THREAD;
+ tc->cmdsize = (uint32_t)(sizeof(struct thread_command)
+ + t->tstate_size);
+ hoffset += sizeof(struct thread_command);
+ /*
+ * Follow with a struct thread_state_flavor and
+ * the appropriate thread state struct for each
+ * thread state flavor.
+ */
+ for (i = 0; i < t->flavor_count; i++) {
+ *(mythread_state_flavor_t *)(header + hoffset) =
+ flavors[i];
+ hoffset += sizeof(mythread_state_flavor_t);
+ thread_getstatus(th_act, flavors[i].flavor,
+ (thread_state_t)(header + hoffset),
+ &flavors[i].count);
+ hoffset += flavors[i].count * sizeof(int);
+ }
- t->hoffset = hoffset;
+ t->hoffset = hoffset;
}
+
/*
- * Create a core image on the file "core".
+ * coredump
+ *
+ * Description: Create a core image on the file "core" for the process
+ * indicated
+ *
+ * Parameters: core_proc Process to dump core [*]
+ * reserve_mb If non-zero, leave filesystem with
+ * at least this much free space.
+ * coredump_flags Extra options (ignore rlimit, run fsync)
+ *
+ * Returns: 0 Success
+ * !0 Failure errno
+ *
+ * IMPORTANT: This function can only be called on the current process, due
+ * to assumptions below; see variable declaration section for
+ * details.
*/
-#define MAX_TSTATE_FLAVORS 10
+#define MAX_TSTATE_FLAVORS 10
int
-coredump(p)
- register struct proc *p;
+coredump(proc_t core_proc, uint32_t reserve_mb, int coredump_flags)
{
- int error=0;
- register struct pcred *pcred = p->p_cred;
- register struct ucred *cred = pcred->pc_ucred;
- struct nameidata nd;
- struct vattr vattr;
- vm_map_t map;
- int thread_count, segment_count;
- int command_size, header_size, tstate_size;
- int hoffset, foffset, vmoffset;
- vm_offset_t header;
- struct machine_slot *ms;
- struct mach_header *mh;
- struct segment_command *sc;
- struct thread_command *tc;
- vm_size_t size;
- vm_prot_t prot;
- vm_prot_t maxprot;
- vm_inherit_t inherit;
- vm_offset_t offset;
- int error1;
- task_t task;
- char core_name[MAXCOMLEN+6];
+/* Begin assumptions that limit us to only the current process */
+ vfs_context_t ctx = vfs_context_current();
+ vm_map_t map = current_map();
+ task_t task = current_task();
+/* End assumptions */
+ kauth_cred_t cred = vfs_context_ucred(ctx);
+ int error = 0;
+ struct vnode_attr va;
+ size_t thread_count, segment_count;
+ size_t command_size, header_size, tstate_size;
+ size_t hoffset;
+ off_t foffset;
+ mach_vm_offset_t vmoffset;
+ vm_offset_t header;
+ mach_vm_size_t vmsize;
+ vm_prot_t prot;
+ vm_prot_t maxprot;
+ vm_inherit_t inherit;
+ int error1 = 0;
+ char stack_name[MAXCOMLEN + 6];
+ char *alloced_name = NULL;
+ char *name = NULL;
mythread_state_flavor_t flavors[MAX_TSTATE_FLAVORS];
- vm_size_t nflavors,mapsize;
- int i;
- int nesting_depth = 0;
- kern_return_t kret;
+ vm_size_t mapsize;
+ size_t i;
+ uint32_t nesting_depth = 0;
+ kern_return_t kret;
struct vm_region_submap_info_64 vbr;
- int vbrcount=0;
+ mach_msg_type_number_t vbrcount = 0;
tir_t tir1;
struct vnode * vp;
- extern boolean_t coredumpok(vm_map_t map, vm_offset_t va); /* temp fix */
- extern task_t current_task(); /* XXX */
+ struct mach_header *mh = NULL; /* protected by is_64 */
+ struct mach_header_64 *mh64 = NULL; /* protected by is_64 */
+ int is_64 = 0;
+ size_t mach_header_sz = sizeof(struct mach_header);
+ size_t segment_command_sz = sizeof(struct segment_command);
+
+ if (current_proc() != core_proc) {
+ panic("coredump() called against proc that is not current_proc: %p", core_proc);
+ }
+
+ if (do_coredump == 0 || /* Not dumping at all */
+ ((sugid_coredump == 0) && /* Not dumping SUID/SGID binaries */
+ ((kauth_cred_getsvuid(cred) != kauth_cred_getruid(cred)) ||
+ (kauth_cred_getsvgid(cred) != kauth_cred_getrgid(cred))))) {
+ error = EFAULT;
+ goto out2;
+ }
+
+#if CONFIG_MACF
+ error = mac_proc_check_dump_core(core_proc);
+ if (error != 0) {
+ goto out2;
+ }
+#endif
- if (pcred->p_svuid != pcred->p_ruid || pcred->p_svgid != pcred->p_rgid)
- return (EFAULT);
+ if (IS_64BIT_PROCESS(core_proc)) {
+ is_64 = 1;
+ mach_header_sz = sizeof(struct mach_header_64);
+ segment_command_sz = sizeof(struct segment_command_64);
+ }
- task = current_task();
- map = current_map();
mapsize = get_vmmap_size(map);
- if (mapsize >= p->p_rlimit[RLIMIT_CORE].rlim_cur)
- return (EFAULT);
- (void) task_suspend(task);
+ if (((coredump_flags & COREDUMP_IGNORE_ULIMIT) == 0) &&
+ (mapsize >= proc_limitgetcur(core_proc, RLIMIT_CORE, FALSE))) {
+ error = EFAULT;
+ goto out2;
+ }
- sprintf(core_name, "/cores/core.%d", p->p_pid);
- NDINIT(&nd, LOOKUP, NOFOLLOW, UIO_SYSSPACE, core_name, p);
- if(error = vn_open(&nd, O_CREAT | FWRITE | O_NOFOLLOW, S_IRUSR ))
- return (error);
- vp = nd.ni_vp;
-
+ (void) task_suspend_internal(task);
+
+ MALLOC(alloced_name, char *, MAXPATHLEN, M_TEMP, M_NOWAIT | M_ZERO);
+
+ /* create name according to sysctl'able format string */
+ /* if name creation fails, fall back to historical behaviour... */
+ if (alloced_name == NULL ||
+ proc_core_name(core_proc->p_comm, kauth_cred_getuid(cred),
+ core_proc->p_pid, alloced_name, MAXPATHLEN)) {
+ snprintf(stack_name, sizeof(stack_name),
+ "/cores/core.%d", core_proc->p_pid);
+ name = stack_name;
+ } else {
+ name = alloced_name;
+ }
+
+ if ((error = vnode_open(name, (O_CREAT | FWRITE | O_NOFOLLOW), S_IRUSR, VNODE_LOOKUP_NOFOLLOW, &vp, ctx))) {
+ goto out2;
+ }
+
+ VATTR_INIT(&va);
+ VATTR_WANTED(&va, va_nlink);
/* Don't dump to non-regular files or files with links. */
if (vp->v_type != VREG ||
- VOP_GETATTR(vp, &vattr, cred, p) || vattr.va_nlink != 1) {
+ vnode_getattr(vp, &va, ctx) || va.va_nlink != 1) {
error = EFAULT;
goto out;
}
- VATTR_NULL(&vattr);
- vattr.va_size = 0;
- VOP_LEASE(vp, p, cred, LEASE_WRITE);
- VOP_SETATTR(vp, &vattr, cred, p);
- p->p_acflag |= ACORE;
+ VATTR_INIT(&va); /* better to do it here than waste more stack in vnode_setsize */
+ VATTR_SET(&va, va_data_size, 0);
+ if (core_proc == initproc) {
+ VATTR_SET(&va, va_dataprotect_class, PROTECTION_CLASS_D);
+ }
+ vnode_setattr(vp, &va, ctx);
+ core_proc->p_acflag |= ACORE;
+
+ if ((reserve_mb > 0) &&
+ ((freespace_mb(vp) - (mapsize >> 20)) < reserve_mb)) {
+ error = ENOSPC;
+ goto out;
+ }
/*
* If the task is modified while dumping the file
*/
thread_count = get_task_numacts(task);
- segment_count = get_vmmap_entries(map); /* XXX */
- /*
- * nflavors here is really the number of ints in flavors
- * to meet the thread_getstatus() calling convention
- */
- nflavors = mynum_flavors;
- bcopy(thread_flavor_array,flavors,sizeof(thread_flavor_array));
+ segment_count = get_vmmap_entries(map); /* XXX */
+ tir1.flavor_count = sizeof(thread_flavor_array) / sizeof(mythread_state_flavor_t);
+ bcopy(thread_flavor_array, flavors, sizeof(thread_flavor_array));
tstate_size = 0;
- for (i = 0; i < nflavors; i++)
+ for (i = 0; i < tir1.flavor_count; i++) {
tstate_size += sizeof(mythread_state_flavor_t) +
- (flavors[i].count * sizeof(int));
+ (flavors[i].count * sizeof(int));
+ }
- command_size = segment_count*sizeof(struct segment_command) +
- thread_count*sizeof(struct thread_command) +
- tstate_size*thread_count;
+ {
+ size_t lhs;
+ size_t rhs;
- header_size = command_size + sizeof(struct mach_header);
+ /* lhs = segment_count * segment_command_sz */
+ if (os_mul_overflow(segment_count, segment_command_sz, &lhs)) {
+ error = ENOMEM;
+ goto out;
+ }
- (void) kmem_alloc_wired(kernel_map,
- (vm_offset_t *)&header,
- (vm_size_t)header_size);
+ /* rhs = (tstate_size + sizeof(struct thread_command)) * thread_count */
+ if (os_add_and_mul_overflow(tstate_size, sizeof(struct thread_command), thread_count, &rhs)) {
+ error = ENOMEM;
+ goto out;
+ }
+
+ /* command_size = lhs + rhs */
+ if (os_add_overflow(lhs, rhs, &command_size)) {
+ error = ENOMEM;
+ goto out;
+ }
+ }
+
+ if (os_add_overflow(command_size, mach_header_sz, &header_size)) {
+ error = ENOMEM;
+ goto out;
+ }
+
+ if (kmem_alloc(kernel_map, &header, (vm_size_t)header_size, VM_KERN_MEMORY_DIAG) != KERN_SUCCESS) {
+ error = ENOMEM;
+ goto out;
+ }
/*
* Set up Mach-O header.
*/
- mh = (struct mach_header *) header;
- ms = &machine_slot[cpu_number()];
- mh->magic = MH_MAGIC;
- mh->cputype = ms->cpu_type;
- mh->cpusubtype = ms->cpu_subtype;
- mh->filetype = MH_CORE;
- mh->ncmds = segment_count + thread_count;
- mh->sizeofcmds = command_size;
-
- hoffset = sizeof(struct mach_header); /* offset into header */
- foffset = round_page_32(header_size); /* offset into file */
- vmoffset = VM_MIN_ADDRESS; /* offset into VM */
+ if (is_64) {
+ mh64 = (struct mach_header_64 *)header;
+ mh64->magic = MH_MAGIC_64;
+ mh64->cputype = process_cpu_type(core_proc);
+ mh64->cpusubtype = process_cpu_subtype(core_proc);
+ mh64->filetype = MH_CORE;
+ mh64->ncmds = (uint32_t)(segment_count + thread_count);
+ mh64->sizeofcmds = (uint32_t)command_size;
+ mh64->reserved = 0; /* 8 byte alignment */
+ } else {
+ mh = (struct mach_header *)header;
+ mh->magic = MH_MAGIC;
+ mh->cputype = process_cpu_type(core_proc);
+ mh->cpusubtype = process_cpu_subtype(core_proc);
+ mh->filetype = MH_CORE;
+ mh->ncmds = (uint32_t)(segment_count + thread_count);
+ mh->sizeofcmds = (uint32_t)command_size;
+ }
+
+ hoffset = mach_header_sz; /* offset into header */
+ foffset = round_page(header_size); /* offset into file */
+ vmoffset = MACH_VM_MIN_ADDRESS; /* offset into VM */
+
/*
- * We use to check for an error, here, now we try and get
+ * We use to check for an error, here, now we try and get
* as much as we can
*/
- while (segment_count > 0){
+ while (segment_count > 0) {
+ struct segment_command *sc;
+ struct segment_command_64 *sc64;
+
/*
* Get region information for next region.
*/
-
+
while (1) {
vbrcount = VM_REGION_SUBMAP_INFO_COUNT_64;
- if((kret = vm_region_recurse_64(map,
- &vmoffset, &size, &nesting_depth,
- &vbr, &vbrcount)) != KERN_SUCCESS) {
+ if ((kret = mach_vm_region_recurse(map,
+ &vmoffset, &vmsize, &nesting_depth,
+ (vm_region_recurse_info_t)&vbr,
+ &vbrcount)) != KERN_SUCCESS) {
break;
}
- if(vbr.is_submap) {
+ /*
+ * If we get a valid mapping back, but we're dumping
+ * a 32 bit process, and it's over the allowable
+ * address space of a 32 bit process, it's the same
+ * as if mach_vm_region_recurse() failed.
+ */
+ if (!(is_64) &&
+ (vmoffset + vmsize > VM_MAX_ADDRESS)) {
+ kret = KERN_INVALID_ADDRESS;
+ break;
+ }
+ if (vbr.is_submap) {
nesting_depth++;
continue;
} else {
break;
}
}
- if(kret != KERN_SUCCESS)
+ if (kret != KERN_SUCCESS) {
break;
+ }
prot = vbr.protection;
maxprot = vbr.max_protection;
/*
* Fill in segment command structure.
*/
- sc = (struct segment_command *) (header + hoffset);
- sc->cmd = LC_SEGMENT;
- sc->cmdsize = sizeof(struct segment_command);
- /* segment name is zerod by kmem_alloc */
- sc->segname[0] = 0;
- sc->vmaddr = vmoffset;
- sc->vmsize = size;
- sc->fileoff = foffset;
- sc->filesize = size;
- sc->maxprot = maxprot;
- sc->initprot = prot;
- sc->nsects = 0;
+ if (is_64) {
+ sc64 = (struct segment_command_64 *)(header + hoffset);
+ sc64->cmd = LC_SEGMENT_64;
+ sc64->cmdsize = sizeof(struct segment_command_64);
+ /* segment name is zeroed by kmem_alloc */
+ sc64->segname[0] = 0;
+ sc64->vmaddr = vmoffset;
+ sc64->vmsize = vmsize;
+ sc64->fileoff = foffset;
+ sc64->filesize = vmsize;
+ sc64->maxprot = maxprot;
+ sc64->initprot = prot;
+ sc64->nsects = 0;
+ sc64->flags = 0;
+ } else {
+ sc = (struct segment_command *) (header + hoffset);
+ sc->cmd = LC_SEGMENT;
+ sc->cmdsize = sizeof(struct segment_command);
+ /* segment name is zeroed by kmem_alloc */
+ sc->segname[0] = 0;
+ sc->vmaddr = CAST_DOWN_EXPLICIT(uint32_t, vmoffset);
+ sc->vmsize = CAST_DOWN_EXPLICIT(uint32_t, vmsize);
+ sc->fileoff = CAST_DOWN_EXPLICIT(uint32_t, foffset); /* will never truncate */
+ sc->filesize = CAST_DOWN_EXPLICIT(uint32_t, vmsize); /* will never truncate */
+ sc->maxprot = maxprot;
+ sc->initprot = prot;
+ sc->nsects = 0;
+ sc->flags = 0;
+ }
/*
* Write segment out. Try as hard as possible to
* get read access to the data.
*/
if ((prot & VM_PROT_READ) == 0) {
- vm_protect(map, vmoffset, size, FALSE,
- prot|VM_PROT_READ);
+ mach_vm_protect(map, vmoffset, vmsize, FALSE,
+ prot | VM_PROT_READ);
}
/*
* Only actually perform write if we can read.
* a hole in the file.
*/
if ((maxprot & VM_PROT_READ) == VM_PROT_READ
- && vbr.user_tag != VM_MEMORY_IOKIT
- && coredumpok(map,vmoffset)) {
- error = vn_rdwr(UIO_WRITE, vp, (caddr_t)vmoffset, size, foffset,
- UIO_USERSPACE, IO_NODELOCKED|IO_UNIT, cred, (int *) 0, p);
+ && vbr.user_tag != VM_MEMORY_IOKIT
+ && coredumpok(map, vmoffset)) {
+ error = vn_rdwr_64(UIO_WRITE, vp, vmoffset, vmsize, foffset,
+ (IS_64BIT_PROCESS(core_proc) ? UIO_USERSPACE64 : UIO_USERSPACE32),
+ IO_NOCACHE | IO_NODELOCKED | IO_UNIT, cred, (int64_t *) 0, core_proc);
}
- hoffset += sizeof(struct segment_command);
- foffset += size;
- vmoffset += size;
+ hoffset += segment_command_sz;
+ foffset += vmsize;
+ vmoffset += vmsize;
segment_count--;
}
+ /*
+ * If there are remaining segments which have not been written
+ * out because break in the loop above, then they were not counted
+ * because they exceed the real address space of the executable
+ * type: remove them from the header's count. This is OK, since
+ * we are allowed to have a sparse area following the segments.
+ */
+ if (is_64) {
+ mh64->ncmds -= segment_count;
+ mh64->sizeofcmds -= segment_count * segment_command_sz;
+ } else {
+ mh->ncmds -= segment_count;
+ mh->sizeofcmds -= segment_count * segment_command_sz;
+ }
+
tir1.header = header;
tir1.hoffset = hoffset;
tir1.flavors = flavors;
tir1.tstate_size = tstate_size;
- task_act_iterate_wth_args(task, collectth_state,&tir1);
+ task_act_iterate_wth_args(task, collectth_state, &tir1);
/*
* Write out the Mach header at the beginning of the
- * file.
+ * file. OK to use a 32 bit write for this.
*/
- error = vn_rdwr(UIO_WRITE, vp, (caddr_t)header, header_size, (off_t)0,
- UIO_SYSSPACE, IO_NODELOCKED|IO_UNIT, cred, (int *) 0, p);
+ error = vn_rdwr(UIO_WRITE, vp, (caddr_t)header, (int)MIN(header_size, INT_MAX), (off_t)0,
+ UIO_SYSSPACE, IO_NOCACHE | IO_NODELOCKED | IO_UNIT, cred, (int *) 0, core_proc);
kmem_free(kernel_map, header, header_size);
+
+ if ((coredump_flags & COREDUMP_FULLFSYNC) && error == 0) {
+ error = VNOP_IOCTL(vp, F_FULLFSYNC, (caddr_t)NULL, 0, ctx);
+ }
out:
- VOP_UNLOCK(vp, 0, p);
- error1 = vn_close(vp, FWRITE, cred, p);
- if (error == 0)
+ error1 = vnode_close(vp, FWRITE, ctx);
+out2:
+#if CONFIG_AUDIT
+ audit_proc_coredump(core_proc, name, error);
+#endif
+ if (alloced_name != NULL) {
+ FREE(alloced_name, M_TEMP);
+ }
+ if (error == 0) {
error = error1;
- return (error);
+ }
+
+ return error;
}
+
+#else /* CONFIG_COREDUMP */
+
+/* When core dumps aren't needed, no need to compile this file at all */
+
+#error assertion failed: this section is not compiled
+
+#endif /* CONFIG_COREDUMP */