/*
- * Copyright (c) 2000 Apple Computer, Inc. All rights reserved.
+ * Copyright (c) 2000-2006 Apple Computer, Inc. All rights reserved.
*
- * @APPLE_LICENSE_HEADER_START@
+ * @APPLE_OSREFERENCE_LICENSE_HEADER_START@
*
- * The contents of this file constitute Original Code as defined in and
- * are subject to the Apple Public Source License Version 1.1 (the
- * "License"). You may not use this file except in compliance with the
- * License. Please obtain a copy of the License at
- * http://www.apple.com/publicsource and read it before using this file.
+ * This file contains Original Code and/or Modifications of Original Code
+ * as defined in and that are subject to the Apple Public Source License
+ * Version 2.0 (the 'License'). You may not use this file except in
+ * compliance with the License. The rights granted to you under the License
+ * may not be used to create, or enable the creation or redistribution of,
+ * unlawful or unlicensed copies of an Apple operating system, or to
+ * circumvent, violate, or enable the circumvention or violation of, any
+ * terms of an Apple operating system software license agreement.
*
- * This Original Code and all software distributed under the License are
- * distributed on an "AS IS" basis, WITHOUT WARRANTY OF ANY KIND, EITHER
+ * Please obtain a copy of the License at
+ * http://www.opensource.apple.com/apsl/ and read it before using this file.
+ *
+ * The Original Code and all software distributed under the License are
+ * distributed on an 'AS IS' basis, WITHOUT WARRANTY OF ANY KIND, EITHER
* EXPRESS OR IMPLIED, AND APPLE HEREBY DISCLAIMS ALL SUCH WARRANTIES,
* INCLUDING WITHOUT LIMITATION, ANY WARRANTIES OF MERCHANTABILITY,
- * FITNESS FOR A PARTICULAR PURPOSE OR NON-INFRINGEMENT. Please see the
- * License for the specific language governing rights and limitations
- * under the License.
+ * FITNESS FOR A PARTICULAR PURPOSE, QUIET ENJOYMENT OR NON-INFRINGEMENT.
+ * Please see the License for the specific language governing rights and
+ * limitations under the License.
*
- * @APPLE_LICENSE_HEADER_END@
+ * @APPLE_OSREFERENCE_LICENSE_HEADER_END@
*/
/* Copyright (c) 1991 NeXT Computer, Inc. All rights reserved.
*
*
* This file contains machine independent code for performing core dumps.
*
- * HISTORY
- * 16-Feb-91 Mike DeMoney (mike@next.com)
- * Massaged into MI form from m68k/core.c.
*/
#include <mach/vm_param.h>
#include <sys/signalvar.h>
#include <sys/resourcevar.h>
#include <sys/namei.h>
-#include <sys/vnode.h>
-#include <sys/proc.h>
+#include <sys/vnode_internal.h>
+#include <sys/proc_internal.h>
+#include <sys/kauth.h>
#include <sys/timeb.h>
#include <sys/times.h>
-#include <sys/buf.h>
#include <sys/acct.h>
-#include <sys/file.h>
+#include <sys/file_internal.h>
#include <sys/uio.h>
#include <sys/kernel.h>
#include <sys/stat.h>
#include <mach-o/loader.h>
#include <mach/vm_region.h>
+#include <mach/vm_statistics.h>
#include <vm/vm_kern.h>
+#include <vm/vm_protos.h> /* last */
+#include <vm/vm_map.h> /* current_map() */
+#include <mach/mach_vm.h> /* mach_vm_region_recurse() */
+#include <mach/task.h> /* task_suspend() */
+#include <kern/task.h> /* get_task_numacts() */
+
+#include <security/audit/audit.h>
typedef struct {
int flavor; /* the number for this flavor */
- int count; /* count of ints in this flavor */
+ mach_msg_type_number_t count; /* count of ints in this flavor */
} mythread_state_flavor_t;
-#if defined (__ppc__)
-
-mythread_state_flavor_t thread_flavor_array[]={
- {PPC_THREAD_STATE , PPC_THREAD_STATE_COUNT},
- {PPC_FLOAT_STATE, PPC_FLOAT_STATE_COUNT},
- {PPC_EXCEPTION_STATE, PPC_EXCEPTION_STATE_COUNT}
- };
-int mynum_flavors=3;
-#elif defined (__i386__)
+#if defined (__i386__) || defined (__x86_64__)
mythread_state_flavor_t thread_flavor_array [] = {
- {i386_THREAD_STATE, i386_THREAD_STATE_COUNT},
- {i386_THREAD_FPSTATE, i386_THREAD_FPSTATE_COUNT},
- {i386_THREAD_EXCEPTSTATE, i386_THREAD_EXCEPTSTATE_COUNT},
- {i386_THREAD_CTHREADSTATE, i386_THREAD_CTHREADSTATE_COUNT},
- {i386_NEW_THREAD_STATE, i386_NEW_THREAD_STATE_COUNT},
- {i386_FLOAT_STATE, i386_FLOAT_STATE_COUNT},
- {i386_ISA_PORT_MAP_STATE, i386_ISA_PORT_MAP_STATE_COUNT},
- {i386_V86_ASSIST_STATE, i386_V86_ASSIST_STATE_COUNT},
- {THREAD_SYSCALL_STATE, i386_THREAD_SYSCALL_STATE_COUNT}
+ {x86_THREAD_STATE, x86_THREAD_STATE_COUNT},
+ {x86_FLOAT_STATE, x86_FLOAT_STATE_COUNT},
+ {x86_EXCEPTION_STATE, x86_EXCEPTION_STATE_COUNT},
};
-int mynum_flavors=9;
-
+int mynum_flavors=3;
#else
#error architecture not supported
#endif
int hoffset;
mythread_state_flavor_t *flavors;
int tstate_size;
+ int flavor_count;
} tir_t;
-collectth_state(thread_act_t th_act, tir_t *t)
+/* XXX should be static */
+void collectth_state(thread_t th_act, void *tirp);
+
+extern int freespace_mb(vnode_t vp);
+
+/* XXX not in a Mach header anywhere */
+kern_return_t thread_getstatus(register thread_t act, int flavor,
+ thread_state_t tstate, mach_msg_type_number_t *count);
+void task_act_iterate_wth_args(task_t, void(*)(thread_t, void *), void *);
+extern kern_return_t task_suspend_internal(task_t);
+
+static cpu_type_t process_cpu_type(proc_t proc);
+static cpu_type_t process_cpu_subtype(proc_t proc);
+
+#ifdef SECURE_KERNEL
+__XNU_PRIVATE_EXTERN int do_coredump = 0; /* default: don't dump cores */
+#else
+__XNU_PRIVATE_EXTERN int do_coredump = 1; /* default: dump cores */
+#endif
+__XNU_PRIVATE_EXTERN int sugid_coredump = 0; /* default: but not SGUID binaries */
+
+
+/* cpu_type returns only the most generic indication of the current CPU. */
+/* in a core we want to know the kind of process. */
+
+static cpu_type_t
+process_cpu_type(proc_t core_proc)
+{
+ cpu_type_t what_we_think;
+#if defined (__i386__) || defined (__x86_64__)
+ if (IS_64BIT_PROCESS(core_proc)) {
+ what_we_think = CPU_TYPE_X86_64;
+ } else {
+ what_we_think = CPU_TYPE_I386;
+ }
+#endif
+ return what_we_think;
+}
+
+static cpu_type_t
+process_cpu_subtype(proc_t core_proc)
+{
+ cpu_type_t what_we_think;
+#if defined (__i386__) || defined (__x86_64__)
+ if (IS_64BIT_PROCESS(core_proc)) {
+ what_we_think = CPU_SUBTYPE_X86_64_ALL;
+ } else {
+ what_we_think = CPU_SUBTYPE_I386_ALL;
+ }
+#endif
+ return what_we_think;
+}
+
+void
+collectth_state(thread_t th_act, void *tirp)
{
vm_offset_t header;
int hoffset, i ;
mythread_state_flavor_t *flavors;
struct thread_command *tc;
+ tir_t *t = (tir_t *)tirp;
+
/*
* Fill in thread command structure.
*/
* the appropriate thread state struct for each
* thread state flavor.
*/
- for (i = 0; i < mynum_flavors; i++) {
+ for (i = 0; i < t->flavor_count; i++) {
*(mythread_state_flavor_t *)(header+hoffset) =
flavors[i];
hoffset += sizeof(mythread_state_flavor_t);
thread_getstatus(th_act, flavors[i].flavor,
- (thread_state_t *)(header+hoffset),
+ (thread_state_t)(header+hoffset),
&flavors[i].count);
hoffset += flavors[i].count*sizeof(int);
}
t->hoffset = hoffset;
}
+
+
/*
- * Create a core image on the file "core".
+ * coredump
+ *
+ * Description: Create a core image on the file "core" for the process
+ * indicated
+ *
+ * Parameters: core_proc Process to dump core [*]
+ * reserve_mb If non-zero, leave filesystem with
+ * at least this much free space.
+ * ignore_ulimit If set, ignore the process's core file ulimit.
+ *
+ * Returns: 0 Success
+ * EFAULT Failed
+ *
+ * IMPORTANT: This function can only be called on the current process, due
+ * to assumptions below; see variable declaration section for
+ * details.
*/
#define MAX_TSTATE_FLAVORS 10
int
-coredump(p)
- register struct proc *p;
+coredump(proc_t core_proc, uint32_t reserve_mb, int ignore_ulimit)
{
- int error=0;
- register struct pcred *pcred = p->p_cred;
- register struct ucred *cred = pcred->pc_ucred;
- struct nameidata nd;
- struct vattr vattr;
- vm_map_t map;
+/* Begin assumptions that limit us to only the current process */
+ vfs_context_t ctx = vfs_context_current();
+ vm_map_t map = current_map();
+ task_t task = current_task();
+/* End assumptions */
+ kauth_cred_t cred = vfs_context_ucred(ctx);
+ int error = 0;
+ struct vnode_attr va;
int thread_count, segment_count;
int command_size, header_size, tstate_size;
- int hoffset, foffset, vmoffset;
+ int hoffset;
+ off_t foffset;
+ mach_vm_offset_t vmoffset;
vm_offset_t header;
- struct machine_slot *ms;
- struct mach_header *mh;
- struct segment_command *sc;
- struct thread_command *tc;
- vm_size_t size;
+ mach_vm_size_t vmsize;
vm_prot_t prot;
vm_prot_t maxprot;
vm_inherit_t inherit;
- vm_offset_t offset;
- int error1;
- task_t task;
- char core_name[MAXCOMLEN+6];
+ int error1 = 0;
+ char stack_name[MAXCOMLEN+6];
+ char *alloced_name = NULL;
+ char *name;
mythread_state_flavor_t flavors[MAX_TSTATE_FLAVORS];
- vm_size_t nflavors,mapsize;
+ vm_size_t mapsize;
int i;
- int nesting_depth = 0;
+ uint32_t nesting_depth = 0;
kern_return_t kret;
struct vm_region_submap_info_64 vbr;
- int vbrcount=0;
+ mach_msg_type_number_t vbrcount = 0;
tir_t tir1;
struct vnode * vp;
+ struct mach_header *mh = NULL; /* protected by is_64 */
+ struct mach_header_64 *mh64 = NULL; /* protected by is_64 */
+ int is_64 = 0;
+ size_t mach_header_sz = sizeof(struct mach_header);
+ size_t segment_command_sz = sizeof(struct segment_command);
+
+ if (current_proc() != core_proc) {
+ panic("coredump() called against proc that is not current_proc: %p", core_proc);
+ }
+ if (do_coredump == 0 || /* Not dumping at all */
+ ( (sugid_coredump == 0) && /* Not dumping SUID/SGID binaries */
+ ( (kauth_cred_getsvuid(cred) != kauth_cred_getruid(cred)) ||
+ (kauth_cred_getsvgid(cred) != kauth_cred_getrgid(cred))))) {
- if (pcred->p_svuid != pcred->p_ruid || pcred->p_svgid != pcred->p_rgid)
+#if CONFIG_AUDIT
+ audit_proc_coredump(core_proc, NULL, EFAULT);
+#endif
return (EFAULT);
+ }
+
+ if (IS_64BIT_PROCESS(core_proc)) {
+ is_64 = 1;
+ mach_header_sz = sizeof(struct mach_header_64);
+ segment_command_sz = sizeof(struct segment_command_64);
+ }
- task = current_task();
- map = current_map();
mapsize = get_vmmap_size(map);
- if (mapsize >= p->p_rlimit[RLIMIT_CORE].rlim_cur)
+ if ((mapsize >= core_proc->p_rlimit[RLIMIT_CORE].rlim_cur) && (ignore_ulimit == 0))
return (EFAULT);
- (void) task_suspend(task);
+ (void) task_suspend_internal(task);
- /*
- * Make sure all registers, etc. are in pcb so they get
- * into core file.
- */
-#if defined (__ppc__)
- fpu_save();
-#endif
- sprintf(core_name, "/cores/core.%d", p->p_pid);
- NDINIT(&nd, LOOKUP, FOLLOW, UIO_SYSSPACE, core_name, p);
- if(error = vn_open(&nd, O_CREAT | FWRITE, S_IRUSR ))
- return (error);
- vp = nd.ni_vp;
-
+ MALLOC(alloced_name, char *, MAXPATHLEN, M_TEMP, M_NOWAIT | M_ZERO);
+
+ /* create name according to sysctl'able format string */
+ /* if name creation fails, fall back to historical behaviour... */
+ if (alloced_name == NULL ||
+ proc_core_name(core_proc->p_comm, kauth_cred_getuid(cred),
+ core_proc->p_pid, alloced_name, MAXPATHLEN)) {
+ snprintf(stack_name, sizeof(stack_name),
+ "/cores/core.%d", core_proc->p_pid);
+ name = stack_name;
+ } else
+ name = alloced_name;
+
+ if ((error = vnode_open(name, (O_CREAT | FWRITE | O_NOFOLLOW), S_IRUSR, VNODE_LOOKUP_NOFOLLOW, &vp, ctx)))
+ goto out2;
+
+ VATTR_INIT(&va);
+ VATTR_WANTED(&va, va_nlink);
/* Don't dump to non-regular files or files with links. */
if (vp->v_type != VREG ||
- VOP_GETATTR(vp, &vattr, cred, p) || vattr.va_nlink != 1) {
+ vnode_getattr(vp, &va, ctx) || va.va_nlink != 1) {
error = EFAULT;
goto out;
}
- VATTR_NULL(&vattr);
- vattr.va_size = 0;
- VOP_LEASE(vp, p, cred, LEASE_WRITE);
- VOP_SETATTR(vp, &vattr, cred, p);
- p->p_acflag |= ACORE;
+ VATTR_INIT(&va); /* better to do it here than waste more stack in vnode_setsize */
+ VATTR_SET(&va, va_data_size, 0);
+ vnode_setattr(vp, &va, ctx);
+ core_proc->p_acflag |= ACORE;
+
+ if ((reserve_mb > 0) &&
+ ((freespace_mb(vp) - (mapsize >> 20)) < reserve_mb)) {
+ error = ENOSPC;
+ goto out;
+ }
/*
* If the task is modified while dumping the file
thread_count = get_task_numacts(task);
segment_count = get_vmmap_entries(map); /* XXX */
- /*
- * nflavors here is really the number of ints in flavors
- * to meet the thread_getstatus() calling convention
- */
-#if 0
- nflavors = sizeof(flavors)/sizeof(int);
- if (thread_getstatus(current_thread(), THREAD_STATE_FLAVOR_LIST,
- (thread_state_t)(flavors),
- &nflavors) != KERN_SUCCESS)
- panic("core flavor list");
- /* now convert to number of flavors */
- nflavors /= sizeof(mythread_state_flavor_t)/sizeof(int);
-#else
- nflavors = mynum_flavors;
- bcopy(thread_flavor_array,flavors,sizeof(thread_flavor_array));
-#endif
+ tir1.flavor_count = sizeof(thread_flavor_array)/sizeof(mythread_state_flavor_t);
+ bcopy(thread_flavor_array, flavors,sizeof(thread_flavor_array));
tstate_size = 0;
- for (i = 0; i < nflavors; i++)
+ for (i = 0; i < tir1.flavor_count; i++)
tstate_size += sizeof(mythread_state_flavor_t) +
(flavors[i].count * sizeof(int));
-
- command_size = segment_count*sizeof(struct segment_command) +
+ command_size = segment_count * segment_command_sz +
thread_count*sizeof(struct thread_command) +
tstate_size*thread_count;
- header_size = command_size + sizeof(struct mach_header);
+ header_size = command_size + mach_header_sz;
- (void) kmem_alloc_wired(kernel_map,
- (vm_offset_t *)&header,
- (vm_size_t)header_size);
+ if (kmem_alloc(kernel_map, &header, (vm_size_t)header_size) != KERN_SUCCESS) {
+ error = ENOMEM;
+ goto out;
+ }
/*
* Set up Mach-O header.
*/
- mh = (struct mach_header *) header;
- ms = &machine_slot[cpu_number()];
- mh->magic = MH_MAGIC;
- mh->cputype = ms->cpu_type;
- mh->cpusubtype = ms->cpu_subtype;
- mh->filetype = MH_CORE;
- mh->ncmds = segment_count + thread_count;
- mh->sizeofcmds = command_size;
-
- hoffset = sizeof(struct mach_header); /* offset into header */
+ if (is_64) {
+ mh64 = (struct mach_header_64 *)header;
+ mh64->magic = MH_MAGIC_64;
+ mh64->cputype = process_cpu_type(core_proc);
+ mh64->cpusubtype = process_cpu_subtype(core_proc);
+ mh64->filetype = MH_CORE;
+ mh64->ncmds = segment_count + thread_count;
+ mh64->sizeofcmds = command_size;
+ mh64->reserved = 0; /* 8 byte alignment */
+ } else {
+ mh = (struct mach_header *)header;
+ mh->magic = MH_MAGIC;
+ mh->cputype = process_cpu_type(core_proc);
+ mh->cpusubtype = process_cpu_subtype(core_proc);
+ mh->filetype = MH_CORE;
+ mh->ncmds = segment_count + thread_count;
+ mh->sizeofcmds = command_size;
+ }
+
+ hoffset = mach_header_sz; /* offset into header */
foffset = round_page(header_size); /* offset into file */
- vmoffset = VM_MIN_ADDRESS; /* offset into VM */
- /* We use to check for an error, here, now we try and get
+ vmoffset = MACH_VM_MIN_ADDRESS; /* offset into VM */
+
+ /*
+ * We use to check for an error, here, now we try and get
* as much as we can
*/
- while (segment_count > 0){
+ while (segment_count > 0) {
+ struct segment_command *sc;
+ struct segment_command_64 *sc64;
+
/*
* Get region information for next region.
*/
while (1) {
vbrcount = VM_REGION_SUBMAP_INFO_COUNT_64;
- if((kret = vm_region_recurse_64(map,
- &vmoffset, &size, &nesting_depth,
- &vbr, &vbrcount)) != KERN_SUCCESS) {
+ if((kret = mach_vm_region_recurse(map,
+ &vmoffset, &vmsize, &nesting_depth,
+ (vm_region_recurse_info_t)&vbr,
+ &vbrcount)) != KERN_SUCCESS) {
+ break;
+ }
+ /*
+ * If we get a valid mapping back, but we're dumping
+ * a 32 bit process, and it's over the allowable
+ * address space of a 32 bit process, it's the same
+ * as if mach_vm_region_recurse() failed.
+ */
+ if (!(is_64) &&
+ (vmoffset + vmsize > VM_MAX_ADDRESS)) {
+ kret = KERN_INVALID_ADDRESS;
break;
}
if(vbr.is_submap) {
/*
* Fill in segment command structure.
*/
- sc = (struct segment_command *) (header + hoffset);
- sc->cmd = LC_SEGMENT;
- sc->cmdsize = sizeof(struct segment_command);
- /* segment name is zerod by kmem_alloc */
- sc->vmaddr = vmoffset;
- sc->vmsize = size;
- sc->fileoff = foffset;
- sc->filesize = size;
- sc->maxprot = maxprot;
- sc->initprot = prot;
- sc->nsects = 0;
+ if (is_64) {
+ sc64 = (struct segment_command_64 *)(header + hoffset);
+ sc64->cmd = LC_SEGMENT_64;
+ sc64->cmdsize = sizeof(struct segment_command_64);
+ /* segment name is zeroed by kmem_alloc */
+ sc64->segname[0] = 0;
+ sc64->vmaddr = vmoffset;
+ sc64->vmsize = vmsize;
+ sc64->fileoff = foffset;
+ sc64->filesize = vmsize;
+ sc64->maxprot = maxprot;
+ sc64->initprot = prot;
+ sc64->nsects = 0;
+ } else {
+ sc = (struct segment_command *) (header + hoffset);
+ sc->cmd = LC_SEGMENT;
+ sc->cmdsize = sizeof(struct segment_command);
+ /* segment name is zeroed by kmem_alloc */
+ sc->segname[0] = 0;
+ sc->vmaddr = CAST_DOWN_EXPLICIT(vm_offset_t,vmoffset);
+ sc->vmsize = CAST_DOWN_EXPLICIT(vm_size_t,vmsize);
+ sc->fileoff = CAST_DOWN_EXPLICIT(uint32_t,foffset); /* will never truncate */
+ sc->filesize = CAST_DOWN_EXPLICIT(uint32_t,vmsize); /* will never truncate */
+ sc->maxprot = maxprot;
+ sc->initprot = prot;
+ sc->nsects = 0;
+ }
/*
* Write segment out. Try as hard as possible to
* get read access to the data.
*/
if ((prot & VM_PROT_READ) == 0) {
- vm_protect(map, vmoffset, size, FALSE,
- prot|VM_PROT_READ);
+ mach_vm_protect(map, vmoffset, vmsize, FALSE,
+ prot|VM_PROT_READ);
}
/*
* Only actually perform write if we can read.
* Note: if we can't read, then we end up with
* a hole in the file.
*/
- if ((maxprot & VM_PROT_READ) == VM_PROT_READ) {
- error = vn_rdwr(UIO_WRITE, vp, (caddr_t)vmoffset, size, foffset,
- UIO_USERSPACE, IO_NODELOCKED|IO_UNIT, cred, (int *) 0, p);
+ if ((maxprot & VM_PROT_READ) == VM_PROT_READ
+ && vbr.user_tag != VM_MEMORY_IOKIT
+ && coredumpok(map,vmoffset)) {
+
+ error = vn_rdwr_64(UIO_WRITE, vp, vmoffset, vmsize, foffset,
+ (IS_64BIT_PROCESS(core_proc) ? UIO_USERSPACE64 : UIO_USERSPACE32),
+ IO_NOCACHE|IO_NODELOCKED|IO_UNIT, cred, (int64_t *) 0, core_proc);
+
}
- hoffset += sizeof(struct segment_command);
- foffset += size;
- vmoffset += size;
+ hoffset += segment_command_sz;
+ foffset += vmsize;
+ vmoffset += vmsize;
segment_count--;
}
-#if 0 /* [ */
- task_lock(task);
- thread = (thread_t) queue_first(&task->thread_list);
- while (thread_count > 0) {
- /*
- * Fill in thread command structure.
- */
- tc = (struct thread_command *) (header + hoffset);
- tc->cmd = LC_THREAD;
- tc->cmdsize = sizeof(struct thread_command)
- + tstate_size;
- hoffset += sizeof(struct thread_command);
- /*
- * Follow with a struct thread_state_flavor and
- * the appropriate thread state struct for each
- * thread state flavor.
- */
- for (i = 0; i < nflavors; i++) {
- *(mythread_state_flavor_t *)(header+hoffset) =
- flavors[i];
- hoffset += sizeof(mythread_state_flavor_t);
- thread_getstatus(thread, flavors[i].flavor,
- (thread_state_t *)(header+hoffset),
- &flavors[i].count);
- hoffset += flavors[i].count*sizeof(int);
- }
- thread = (thread_t) queue_next(&thread->thread_list);
- thread_count--;
+ /*
+ * If there are remaining segments which have not been written
+ * out because break in the loop above, then they were not counted
+ * because they exceed the real address space of the executable
+ * type: remove them from the header's count. This is OK, since
+ * we are allowed to have a sparse area following the segments.
+ */
+ if (is_64) {
+ mh64->ncmds -= segment_count;
+ mh64->sizeofcmds -= segment_count * segment_command_sz;
+ } else {
+ mh->ncmds -= segment_count;
+ mh->sizeofcmds -= segment_count * segment_command_sz;
}
- task_unlock(task);
-#else /* /* 0 ][ */
+
tir1.header = header;
tir1.hoffset = hoffset;
tir1.flavors = flavors;
tir1.tstate_size = tstate_size;
task_act_iterate_wth_args(task, collectth_state,&tir1);
-#endif /* 0 ] */
/*
* Write out the Mach header at the beginning of the
- * file.
+ * file. OK to use a 32 bit write for this.
*/
error = vn_rdwr(UIO_WRITE, vp, (caddr_t)header, header_size, (off_t)0,
- UIO_SYSSPACE, IO_NODELOCKED|IO_UNIT, cred, (int *) 0, p);
+ UIO_SYSSPACE, IO_NOCACHE|IO_NODELOCKED|IO_UNIT, cred, (int *) 0, core_proc);
kmem_free(kernel_map, header, header_size);
out:
- VOP_UNLOCK(vp, 0, p);
- error1 = vn_close(vp, FWRITE, cred, p);
+ error1 = vnode_close(vp, FWRITE, ctx);
+out2:
+#if CONFIG_AUDIT
+ audit_proc_coredump(core_proc, name, error);
+#endif
+ if (alloced_name != NULL)
+ FREE(alloced_name, M_TEMP);
if (error == 0)
error = error1;
+
+ return (error);
}