]> git.saurik.com Git - apple/xnu.git/blobdiff - libsyscall/mach/mach_vm.c
xnu-7195.81.3.tar.gz
[apple/xnu.git] / libsyscall / mach / mach_vm.c
index 2db3830214b03dbd8fcf004b1913ca9d3fb8e402..8b8dfa3fa4374fc78184baafb6af7ec2602a4c5c 100644 (file)
@@ -2,7 +2,7 @@
  * Copyright (c) 2011 Apple Inc. All rights reserved.
  *
  * @APPLE_OSREFERENCE_LICENSE_HEADER_START@
  * Copyright (c) 2011 Apple Inc. All rights reserved.
  *
  * @APPLE_OSREFERENCE_LICENSE_HEADER_START@
- * 
+ *
  * This file contains Original Code and/or Modifications of Original Code
  * as defined in and that are subject to the Apple Public Source License
  * Version 2.0 (the 'License'). You may not use this file except in
  * This file contains Original Code and/or Modifications of Original Code
  * as defined in and that are subject to the Apple Public Source License
  * Version 2.0 (the 'License'). You may not use this file except in
  * unlawful or unlicensed copies of an Apple operating system, or to
  * circumvent, violate, or enable the circumvention or violation of, any
  * terms of an Apple operating system software license agreement.
  * unlawful or unlicensed copies of an Apple operating system, or to
  * circumvent, violate, or enable the circumvention or violation of, any
  * terms of an Apple operating system software license agreement.
- * 
+ *
  * Please obtain a copy of the License at
  * http://www.opensource.apple.com/apsl/ and read it before using this file.
  * Please obtain a copy of the License at
  * http://www.opensource.apple.com/apsl/ and read it before using this file.
- * 
+ *
  * The Original Code and all software distributed under the License are
  * distributed on an 'AS IS' basis, WITHOUT WARRANTY OF ANY KIND, EITHER
  * EXPRESS OR IMPLIED, AND APPLE HEREBY DISCLAIMS ALL SUCH WARRANTIES,
  * The Original Code and all software distributed under the License are
  * distributed on an 'AS IS' basis, WITHOUT WARRANTY OF ANY KIND, EITHER
  * EXPRESS OR IMPLIED, AND APPLE HEREBY DISCLAIMS ALL SUCH WARRANTIES,
@@ -22,7 +22,7 @@
  * FITNESS FOR A PARTICULAR PURPOSE, QUIET ENJOYMENT OR NON-INFRINGEMENT.
  * Please see the License for the specific language governing rights and
  * limitations under the License.
  * FITNESS FOR A PARTICULAR PURPOSE, QUIET ENJOYMENT OR NON-INFRINGEMENT.
  * Please see the License for the specific language governing rights and
  * limitations under the License.
- * 
+ *
  * @APPLE_OSREFERENCE_LICENSE_HEADER_END@
  */
 
  * @APPLE_OSREFERENCE_LICENSE_HEADER_END@
  */
 
 #undef _mach_vm_user_
 #include <mach/mach_vm_internal.h>
 
 #undef _mach_vm_user_
 #include <mach/mach_vm_internal.h>
 
+#include "stack_logging_internal.h"
+
+malloc_logger_t *__syscall_logger = NULL;   // This may get set by Libc's malloc stack logging initialization code.
+
 kern_return_t
 mach_vm_allocate(
 kern_return_t
 mach_vm_allocate(
-               mach_port_name_t target,
-               mach_vm_address_t *address,
-               mach_vm_size_t size,
-               int flags)
+       mach_port_name_t target,
+       mach_vm_address_t *address,
+       mach_vm_size_t size,
+       int flags)
 {
        kern_return_t rv;
 
        rv = _kernelrpc_mach_vm_allocate_trap(target, address, size, flags);
 
 {
        kern_return_t rv;
 
        rv = _kernelrpc_mach_vm_allocate_trap(target, address, size, flags);
 
-       if (rv == MACH_SEND_INVALID_DEST)
+       if (rv == MACH_SEND_INVALID_DEST) {
                rv = _kernelrpc_mach_vm_allocate(target, address, size, flags);
                rv = _kernelrpc_mach_vm_allocate(target, address, size, flags);
+       }
 
 
-       return (rv);
+       int userTagFlags = flags & VM_FLAGS_ALIAS_MASK;
+       if (__syscall_logger && rv == KERN_SUCCESS && (userTagFlags != VM_MAKE_TAG(VM_MEMORY_STACK))) {
+               __syscall_logger(stack_logging_type_vm_allocate | userTagFlags, (uintptr_t)target, (uintptr_t)size, 0, (uintptr_t)*address, 0);
+       }
+
+       return rv;
 }
 
 kern_return_t
 }
 
 kern_return_t
@@ -64,12 +74,17 @@ mach_vm_deallocate(
 {
        kern_return_t rv;
 
 {
        kern_return_t rv;
 
+       if (__syscall_logger) {
+               __syscall_logger(stack_logging_type_vm_deallocate, (uintptr_t)target, (uintptr_t)address, (uintptr_t)size, 0, 0);
+       }
+
        rv = _kernelrpc_mach_vm_deallocate_trap(target, address, size);
 
        rv = _kernelrpc_mach_vm_deallocate_trap(target, address, size);
 
-       if (rv == MACH_SEND_INVALID_DEST)
+       if (rv == MACH_SEND_INVALID_DEST) {
                rv = _kernelrpc_mach_vm_deallocate(target, address, size);
                rv = _kernelrpc_mach_vm_deallocate(target, address, size);
+       }
 
 
-       return (rv);
+       return rv;
 }
 
 kern_return_t
 }
 
 kern_return_t
@@ -83,13 +98,14 @@ mach_vm_protect(
        kern_return_t rv;
 
        rv = _kernelrpc_mach_vm_protect_trap(task, address, size, set_maximum,
        kern_return_t rv;
 
        rv = _kernelrpc_mach_vm_protect_trap(task, address, size, set_maximum,
-               new_protection);
+           new_protection);
 
 
-       if (rv == MACH_SEND_INVALID_DEST)
+       if (rv == MACH_SEND_INVALID_DEST) {
                rv = _kernelrpc_mach_vm_protect(task, address, size,
                rv = _kernelrpc_mach_vm_protect(task, address, size,
-                       set_maximum, new_protection);
+                   set_maximum, new_protection);
+       }
 
 
-       return (rv);
+       return rv;
 }
 
 kern_return_t
 }
 
 kern_return_t
@@ -110,7 +126,7 @@ vm_allocate(
        *address = (vm_address_t)(mach_addr & ((vm_address_t)-1));
 #endif
 
        *address = (vm_address_t)(mach_addr & ((vm_address_t)-1));
 #endif
 
-       return (rv);
+       return rv;
 }
 
 kern_return_t
 }
 
 kern_return_t
@@ -123,7 +139,7 @@ vm_deallocate(
 
        rv = mach_vm_deallocate(task, address, size);
 
 
        rv = mach_vm_deallocate(task, address, size);
 
-       return (rv);
+       return rv;
 }
 
 kern_return_t
 }
 
 kern_return_t
@@ -138,5 +154,202 @@ vm_protect(
 
        rv = mach_vm_protect(task, address, size, set_maximum, new_protection);
 
 
        rv = mach_vm_protect(task, address, size, set_maximum, new_protection);
 
-       return (rv);
+       return rv;
+}
+
+kern_return_t
+mach_vm_map(
+       mach_port_name_t target,
+       mach_vm_address_t *address,
+       mach_vm_size_t size,
+       mach_vm_offset_t mask,
+       int flags,
+       mem_entry_name_port_t object,
+       memory_object_offset_t offset,
+       boolean_t copy,
+       vm_prot_t cur_protection,
+       vm_prot_t max_protection,
+       vm_inherit_t inheritance)
+{
+       kern_return_t rv = MACH_SEND_INVALID_DEST;
+
+       if (object == MEMORY_OBJECT_NULL && max_protection == VM_PROT_ALL &&
+           inheritance == VM_INHERIT_DEFAULT) {
+               rv = _kernelrpc_mach_vm_map_trap(target, address, size, mask, flags,
+                   cur_protection);
+       }
+
+       if (rv == MACH_SEND_INVALID_DEST) {
+               rv = _kernelrpc_mach_vm_map(target, address, size, mask, flags, object,
+                   offset, copy, cur_protection, max_protection, inheritance);
+       }
+
+       int userTagFlags = flags & VM_FLAGS_ALIAS_MASK;
+       if (__syscall_logger && rv == KERN_SUCCESS && (userTagFlags != VM_MAKE_TAG(VM_MEMORY_STACK))) {
+               int eventTypeFlags = stack_logging_type_vm_allocate | stack_logging_type_mapped_file_or_shared_mem;
+               __syscall_logger(eventTypeFlags | userTagFlags, (uintptr_t)target, (uintptr_t)size, 0, (uintptr_t)*address, 0);
+       }
+
+       return rv;
+}
+
+kern_return_t
+mach_vm_remap(
+       mach_port_name_t target,
+       mach_vm_address_t *address,
+       mach_vm_size_t size,
+       mach_vm_offset_t mask,
+       int flags,
+       mach_port_name_t src_task,
+       mach_vm_address_t src_address,
+       boolean_t copy,
+       vm_prot_t *cur_protection,
+       vm_prot_t *max_protection,
+       vm_inherit_t inheritance)
+{
+       kern_return_t rv;
+
+       rv = _kernelrpc_mach_vm_remap(target, address, size, mask, flags,
+           src_task, src_address, copy, cur_protection, max_protection,
+           inheritance);
+
+       if (__syscall_logger && rv == KERN_SUCCESS) {
+               int eventTypeFlags = stack_logging_type_vm_allocate | stack_logging_type_mapped_file_or_shared_mem;
+               int userTagFlags = flags & VM_FLAGS_ALIAS_MASK;
+               __syscall_logger(eventTypeFlags | userTagFlags, (uintptr_t)target, (uintptr_t)size, 0, (uintptr_t)*address, 0);
+       }
+
+       return rv;
+}
+
+kern_return_t
+mach_vm_read(
+       mach_port_name_t target,
+       mach_vm_address_t address,
+       mach_vm_size_t size,
+       vm_offset_t *data,
+       mach_msg_type_number_t *dataCnt)
+{
+       kern_return_t rv;
+
+       rv = _kernelrpc_mach_vm_read(target, address, size, data, dataCnt);
+
+       if (__syscall_logger && rv == KERN_SUCCESS) {
+               int eventTypeFlags = stack_logging_type_vm_allocate | stack_logging_type_mapped_file_or_shared_mem;
+               // The target argument is the remote task from which data is being read,
+               // so pass mach_task_self() as the destination task receiving the allocation.
+               __syscall_logger(eventTypeFlags, (uintptr_t)mach_task_self(), (uintptr_t)*dataCnt, 0, *data, 0);
+       }
+
+       return rv;
+}
+
+kern_return_t
+vm_map(
+       mach_port_name_t target,
+       vm_address_t *address,
+       vm_size_t size,
+       vm_offset_t mask,
+       int flags,
+       mem_entry_name_port_t object,
+       vm_offset_t offset,
+       boolean_t copy,
+       vm_prot_t cur_protection,
+       vm_prot_t max_protection,
+       vm_inherit_t inheritance)
+{
+       kern_return_t rv;
+
+       rv = _kernelrpc_vm_map(target, address, size, mask, flags, object,
+           offset, copy, cur_protection, max_protection, inheritance);
+
+       if (__syscall_logger && rv == KERN_SUCCESS) {
+               int eventTypeFlags = stack_logging_type_vm_allocate | stack_logging_type_mapped_file_or_shared_mem;
+               int userTagFlags = flags & VM_FLAGS_ALIAS_MASK;
+               __syscall_logger(eventTypeFlags | userTagFlags, (uintptr_t)target, (uintptr_t)size, 0, (uintptr_t)*address, 0);
+       }
+
+       return rv;
+}
+
+kern_return_t
+vm_remap(
+       mach_port_name_t target,
+       vm_address_t *address,
+       vm_size_t size,
+       vm_offset_t mask,
+       int flags,
+       mach_port_name_t src_task,
+       vm_address_t src_address,
+       boolean_t copy,
+       vm_prot_t *cur_protection,
+       vm_prot_t *max_protection,
+       vm_inherit_t inheritance)
+{
+       kern_return_t rv;
+
+       rv = _kernelrpc_vm_remap(target, address, size, mask, flags,
+           src_task, src_address, copy, cur_protection, max_protection,
+           inheritance);
+
+       if (__syscall_logger) {
+               int eventTypeFlags = stack_logging_type_vm_allocate | stack_logging_type_mapped_file_or_shared_mem;
+               int userTagFlags = flags & VM_FLAGS_ALIAS_MASK;
+               __syscall_logger(eventTypeFlags | userTagFlags, (uintptr_t)target, (uintptr_t)size, 0, (uintptr_t)*address, 0);
+       }
+
+       return rv;
+}
+
+kern_return_t
+vm_read(
+       mach_port_name_t target,
+       vm_address_t address,
+       vm_size_t size,
+       vm_offset_t *data,
+       mach_msg_type_number_t *dataCnt)
+{
+       kern_return_t rv;
+
+       rv = _kernelrpc_vm_read(target, address, size, data, dataCnt);
+
+       if (__syscall_logger && rv == KERN_SUCCESS) {
+               int eventTypeFlags = stack_logging_type_vm_allocate | stack_logging_type_mapped_file_or_shared_mem;
+               // The target argument is the remote task from which data is being read,
+               // so pass mach_task_self() as the destination task receiving the allocation.
+               __syscall_logger(eventTypeFlags, (uintptr_t)mach_task_self(), (uintptr_t)*dataCnt, 0, *data, 0);
+       }
+
+       return rv;
+}
+
+kern_return_t
+mach_vm_purgable_control(
+       mach_port_name_t        target,
+       mach_vm_offset_t        address,
+       vm_purgable_t           control,
+       int                     *state)
+{
+       kern_return_t rv;
+
+       rv = _kernelrpc_mach_vm_purgable_control_trap(target, address, control, state);
+
+       if (rv == MACH_SEND_INVALID_DEST) {
+               rv = _kernelrpc_mach_vm_purgable_control(target, address, control, state);
+       }
+
+       return rv;
+}
+
+kern_return_t
+vm_purgable_control(
+       mach_port_name_t        task,
+       vm_offset_t             address,
+       vm_purgable_t           control,
+       int                     *state)
+{
+       return mach_vm_purgable_control(task,
+                  (mach_vm_offset_t) address,
+                  control,
+                  state);
 }
 }