X-Git-Url: https://git.saurik.com/apple/xnu.git/blobdiff_plain/2d21ac55c334faf3a56e5634905ed6987fc787d4..ea3f04195ba4a5034c9c8e9b726d4f7ce96f1832:/osfmk/vm/vm_kern.h?ds=inline diff --git a/osfmk/vm/vm_kern.h b/osfmk/vm/vm_kern.h index 3e22b0633..7a0466951 100644 --- a/osfmk/vm/vm_kern.h +++ b/osfmk/vm/vm_kern.h @@ -2,7 +2,7 @@ * Copyright (c) 2000-2006 Apple Computer, Inc. All rights reserved. * * @APPLE_OSREFERENCE_LICENSE_HEADER_START@ - * + * * This file contains Original Code and/or Modifications of Original Code * as defined in and that are subject to the Apple Public Source License * Version 2.0 (the 'License'). You may not use this file except in @@ -11,10 +11,10 @@ * unlawful or unlicensed copies of an Apple operating system, or to * circumvent, violate, or enable the circumvention or violation of, any * terms of an Apple operating system software license agreement. - * + * * Please obtain a copy of the License at * http://www.opensource.apple.com/apsl/ and read it before using this file. - * + * * The Original Code and all software distributed under the License are * distributed on an 'AS IS' basis, WITHOUT WARRANTY OF ANY KIND, EITHER * EXPRESS OR IMPLIED, AND APPLE HEREBY DISCLAIMS ALL SUCH WARRANTIES, @@ -22,34 +22,34 @@ * FITNESS FOR A PARTICULAR PURPOSE, QUIET ENJOYMENT OR NON-INFRINGEMENT. * Please see the License for the specific language governing rights and * limitations under the License. - * + * * @APPLE_OSREFERENCE_LICENSE_HEADER_END@ */ /* * @OSF_COPYRIGHT@ */ -/* +/* * Mach Operating System * Copyright (c) 1991,1990,1989,1988,1987 Carnegie Mellon University * All Rights Reserved. - * + * * Permission to use, copy, modify and distribute this software and its * documentation is hereby granted, provided that both the copyright * notice and this permission notice appear in all copies of the * software, derivative works or modified versions, and any portions * thereof, and that both notices appear in supporting documentation. - * + * * CARNEGIE MELLON ALLOWS FREE USE OF THIS SOFTWARE IN ITS "AS IS" * CONDITION. CARNEGIE MELLON DISCLAIMS ANY LIABILITY OF ANY KIND FOR * ANY DAMAGES WHATSOEVER RESULTING FROM THE USE OF THIS SOFTWARE. - * + * * Carnegie Mellon requests users of this software to return to - * + * * Software Distribution Coordinator or Software.Distribution@CS.CMU.EDU * School of Computer Science * Carnegie Mellon University * Pittsburgh PA 15213-3890 - * + * * any improvements or extensions that they make and grant Carnegie Mellon * the rights to redistribute these changes. */ @@ -63,113 +63,413 @@ * Kernel memory management definitions. */ -#ifndef _VM_VM_KERN_H_ +#ifndef _VM_VM_KERN_H_ #define _VM_VM_KERN_H_ +#ifdef __cplusplus +extern "C" { +#endif + #include #include #include -#ifdef KERNEL_PRIVATE +#ifdef KERNEL_PRIVATE + +#ifdef XNU_KERNEL_PRIVATE -extern kern_return_t kernel_memory_allocate( - vm_map_t map, - vm_offset_t *addrp, - vm_size_t size, - vm_offset_t mask, - int flags); +#include + +extern kern_return_t kernel_memory_allocate( + vm_map_t map, + vm_offset_t *addrp, + vm_size_t size, + vm_offset_t mask, + int flags, + vm_tag_t tag); /* flags for kernel_memory_allocate */ -#define KMA_HERE 0x01 -#define KMA_NOPAGEWAIT 0x02 -#define KMA_KOBJECT 0x04 -#define KMA_LOMEM 0x08 -#define KMA_GUARD_FIRST 0x10 -#define KMA_GUARD_LAST 0x20 +#define KMA_HERE 0x01 +#define KMA_NOPAGEWAIT 0x02 +#define KMA_KOBJECT 0x04 +#define KMA_LOMEM 0x08 +#define KMA_GUARD_FIRST 0x10 +#define KMA_GUARD_LAST 0x20 +#define KMA_PERMANENT 0x40 +#define KMA_NOENCRYPT 0x80 +#define KMA_KSTACK 0x100 +#define KMA_VAONLY 0x200 +#define KMA_COMPRESSOR 0x400 /* Pages belonging to the compressor are not on the paging queues, nor are they counted as wired. */ +#define KMA_ATOMIC 0x800 +#define KMA_ZERO 0x1000 +#define KMA_PAGEABLE 0x2000 + +extern kern_return_t kmem_alloc( + vm_map_t map, + vm_offset_t *addrp, + vm_size_t size, + vm_tag_t tag) __XNU_INTERNAL(kmem_alloc); extern kern_return_t kmem_alloc_contig( - vm_map_t map, - vm_offset_t *addrp, - vm_size_t size, - vm_offset_t mask, - ppnum_t max_pnum, - int flags); - -extern kern_return_t kmem_alloc( - vm_map_t map, - vm_offset_t *addrp, - vm_size_t size); - -extern kern_return_t kmem_alloc_pageable( - vm_map_t map, - vm_offset_t *addrp, - vm_size_t size); - -extern kern_return_t kmem_alloc_aligned( - vm_map_t map, - vm_offset_t *addrp, - vm_size_t size); - -extern kern_return_t kmem_realloc( - vm_map_t map, - vm_offset_t oldaddr, - vm_size_t oldsize, - vm_offset_t *newaddrp, - vm_size_t newsize); - -extern void kmem_free( - vm_map_t map, - vm_offset_t addr, - vm_size_t size); - -extern kern_return_t kmem_suballoc( - vm_map_t parent, - vm_offset_t *addr, - vm_size_t size, - boolean_t pageable, - boolean_t anywhere, - vm_map_t *new_map); + vm_map_t map, + vm_offset_t *addrp, + vm_size_t size, + vm_offset_t mask, + ppnum_t max_pnum, + ppnum_t pnum_mask, + int flags, + vm_tag_t tag); + +extern kern_return_t kmem_alloc_flags( + vm_map_t map, + vm_offset_t *addrp, + vm_size_t size, + vm_tag_t tag, + int flags); + +extern kern_return_t kmem_alloc_pageable( + vm_map_t map, + vm_offset_t *addrp, + vm_size_t size, + vm_tag_t tag) __XNU_INTERNAL(kmem_alloc_pageable); + +extern kern_return_t kmem_alloc_aligned( + vm_map_t map, + vm_offset_t *addrp, + vm_size_t size, + vm_tag_t tag); + +extern kern_return_t kmem_realloc( + vm_map_t map, + vm_offset_t oldaddr, + vm_size_t oldsize, + vm_offset_t *newaddrp, + vm_size_t newsize, + vm_tag_t tag); + +extern void kmem_free( + vm_map_t map, + vm_offset_t addr, + vm_size_t size); + +extern kern_return_t kmem_suballoc( + vm_map_t parent, + vm_offset_t *addr, + vm_size_t size, + boolean_t pageable, + int flags, + vm_map_kernel_flags_t vmk_flags, + vm_tag_t tag, + vm_map_t *new_map); + +extern kern_return_t kmem_alloc_kobject( + vm_map_t map, + vm_offset_t *addrp, + vm_size_t size, + vm_tag_t tag) __XNU_INTERNAL(kmem_alloc_kobject); + +extern kern_return_t kernel_memory_populate( + vm_map_t map, + vm_offset_t addr, + vm_size_t size, + int flags, + vm_tag_t tag); + +extern void kernel_memory_depopulate( + vm_map_t map, + vm_offset_t addr, + vm_size_t size, + int flags); + +extern kern_return_t memory_object_iopl_request( + ipc_port_t port, + memory_object_offset_t offset, + upl_size_t *upl_size, + upl_t *upl_ptr, + upl_page_info_array_t user_page_list, + unsigned int *page_list_count, + upl_control_flags_t *flags, + vm_tag_t tag); + +struct mach_memory_info; +extern kern_return_t vm_page_diagnose(struct mach_memory_info * info, + unsigned int num_info, uint64_t zones_collectable_bytes); + +extern uint32_t vm_page_diagnose_estimate(void); + +#if DEBUG || DEVELOPMENT + +extern kern_return_t mach_memory_info_check(void); + +extern kern_return_t vm_kern_allocation_info(uintptr_t addr, vm_size_t * size, vm_tag_t * tag, vm_size_t * zone_size); + +#endif /* DEBUG || DEVELOPMENT */ + +extern vm_tag_t vm_tag_bt(void); + +extern vm_tag_t vm_tag_alloc(vm_allocation_site_t * site); + +extern void vm_tag_alloc_locked(vm_allocation_site_t * site, vm_allocation_site_t ** releasesiteP); + +extern void vm_tag_update_size(vm_tag_t tag, int64_t size); + +#if VM_MAX_TAG_ZONES +extern void vm_allocation_zones_init(void); +extern void vm_tag_will_update_zone(vm_tag_t tag, uint32_t zidx); +extern void vm_tag_update_zone_size(vm_tag_t tag, uint32_t zidx, int64_t delta, int64_t dwaste); + +extern vm_allocation_zone_total_t ** vm_allocation_zone_totals; + +#endif /* VM_MAX_TAG_ZONES */ + +extern vm_tag_t vm_tag_bt_debug(void); + +extern uint32_t vm_tag_get_kext(vm_tag_t tag, char * name, vm_size_t namelen); + +extern boolean_t vm_kernel_map_is_kernel(vm_map_t map); + +extern ppnum_t kernel_pmap_present_mapping(uint64_t vaddr, uint64_t * pvincr, uintptr_t * pvphysaddr); + +#else /* XNU_KERNEL_PRIVATE */ + +extern kern_return_t kmem_alloc( + vm_map_t map, + vm_offset_t *addrp, + vm_size_t size); + +extern kern_return_t kmem_alloc_pageable( + vm_map_t map, + vm_offset_t *addrp, + vm_size_t size); + +extern kern_return_t kmem_alloc_kobject( + vm_map_t map, + vm_offset_t *addrp, + vm_size_t size); + +extern void kmem_free( + vm_map_t map, + vm_offset_t addr, + vm_size_t size); + +#endif /* !XNU_KERNEL_PRIVATE */ #ifdef XNU_KERNEL_PRIVATE -extern kern_return_t kmem_alloc_wired( - vm_map_t map, - vm_offset_t *addrp, - vm_size_t size); -#endif +typedef struct vm_allocation_site kern_allocation_name; +typedef kern_allocation_name * kern_allocation_name_t; +#else /* XNU_KERNEL_PRIVATE */ +struct kern_allocation_name; +typedef struct kern_allocation_name * kern_allocation_name_t; +#endif /* !XNU_KERNEL_PRIVATE */ + +extern kern_allocation_name_t kern_allocation_name_allocate(const char * name, uint32_t suballocs); +extern void kern_allocation_name_release(kern_allocation_name_t allocation); +extern const char * kern_allocation_get_name(kern_allocation_name_t allocation); +#ifdef XNU_KERNEL_PRIVATE +extern void kern_allocation_update_size(kern_allocation_name_t allocation, int64_t delta); +extern void kern_allocation_update_subtotal(kern_allocation_name_t allocation, uint32_t subtag, int64_t delta); +extern vm_tag_t kern_allocation_name_get_vm_tag(kern_allocation_name_t allocation); +#endif /* XNU_KERNEL_PRIVATE */ + +#ifdef MACH_KERNEL_PRIVATE + +extern void kmem_init( + vm_offset_t start, + vm_offset_t end); + +extern kern_return_t copyinmap( + vm_map_t map, + vm_map_offset_t fromaddr, + void *todata, + vm_size_t length); + +extern kern_return_t copyoutmap( + vm_map_t map, + void *fromdata, + vm_map_offset_t toaddr, + vm_size_t length); + +extern kern_return_t kmem_alloc_external( + vm_map_t map, + vm_offset_t *addrp, + vm_size_t size); + +extern kern_return_t kmem_alloc_kobject_external( + vm_map_t map, + vm_offset_t *addrp, + vm_size_t size); + +extern kern_return_t kmem_alloc_pageable_external( + vm_map_t map, + vm_offset_t *addrp, + vm_size_t size); + +#endif /* MACH_KERNEL_PRIVATE */ + +#ifdef XNU_KERNEL_PRIVATE + +extern kern_return_t mach_vm_allocate_kernel( + vm_map_t map, + mach_vm_offset_t *addr, + mach_vm_size_t size, + int flags, + vm_tag_t tag); + +extern kern_return_t vm_allocate_kernel( + vm_map_t map, + vm_offset_t *addr, + vm_size_t size, + int flags, + vm_tag_t tag); + + +extern kern_return_t mach_vm_map_kernel( + vm_map_t target_map, + mach_vm_offset_t *address, + mach_vm_size_t initial_size, + mach_vm_offset_t mask, + int flags, + vm_map_kernel_flags_t vmk_flags, + vm_tag_t tag, + ipc_port_t port, + vm_object_offset_t offset, + boolean_t copy, + vm_prot_t cur_protection, + vm_prot_t max_protection, + vm_inherit_t inheritance); -#ifdef MACH_KERNEL_PRIVATE -extern void kmem_init( - vm_offset_t start, - vm_offset_t end) __attribute__((section("__TEXT, initcode"))); +extern kern_return_t vm_map_kernel( + vm_map_t target_map, + vm_offset_t *address, + vm_size_t size, + vm_offset_t mask, + int flags, + vm_map_kernel_flags_t vmk_flags, + vm_tag_t tag, + ipc_port_t port, + vm_offset_t offset, + boolean_t copy, + vm_prot_t cur_protection, + vm_prot_t max_protection, + vm_inherit_t inheritance); +extern kern_return_t mach_vm_remap_kernel( + vm_map_t target_map, + mach_vm_offset_t *address, + mach_vm_size_t size, + mach_vm_offset_t mask, + int flags, + vm_tag_t tag, + vm_map_t src_map, + mach_vm_offset_t memory_address, + boolean_t copy, + vm_prot_t *cur_protection, + vm_prot_t *max_protection, + vm_inherit_t inheritance); -extern kern_return_t copyinmap( - vm_map_t map, - vm_map_offset_t fromaddr, - void *todata, - vm_size_t length); +extern kern_return_t vm_remap_kernel( + vm_map_t target_map, + vm_offset_t *address, + vm_size_t size, + vm_offset_t mask, + int flags, + vm_tag_t tag, + vm_map_t src_map, + vm_offset_t memory_address, + boolean_t copy, + vm_prot_t *cur_protection, + vm_prot_t *max_protection, + vm_inherit_t inheritance); -extern kern_return_t copyoutmap( - vm_map_t map, - void *fromdata, - vm_map_offset_t toaddr, - vm_size_t length); +extern kern_return_t vm_map_64_kernel( + vm_map_t target_map, + vm_offset_t *address, + vm_size_t size, + vm_offset_t mask, + int flags, + vm_map_kernel_flags_t vmk_flags, + vm_tag_t tag, + ipc_port_t port, + vm_object_offset_t offset, + boolean_t copy, + vm_prot_t cur_protection, + vm_prot_t max_protection, + vm_inherit_t inheritance); -extern kern_return_t vm_conflict_check( - vm_map_t map, - vm_map_offset_t off, - vm_map_size_t len, - memory_object_t pager, - vm_object_offset_t file_off); +extern kern_return_t mach_vm_wire_kernel( + host_priv_t host_priv, + vm_map_t map, + mach_vm_offset_t start, + mach_vm_size_t size, + vm_prot_t access, + vm_tag_t tag); -#endif /* MACH_KERNEL_PRIVATE */ +extern kern_return_t vm_map_wire_kernel( + vm_map_t map, + vm_map_offset_t start, + vm_map_offset_t end, + vm_prot_t caller_prot, + vm_tag_t tag, + boolean_t user_wire); -extern vm_map_t kernel_map; -extern vm_map_t kernel_pageable_map; +extern kern_return_t vm_map_wire_and_extract_kernel( + vm_map_t map, + vm_map_offset_t start, + vm_prot_t caller_prot, + vm_tag_t tag, + boolean_t user_wire, + ppnum_t *physpage_p); + +#endif /* XNU_KERNEL_PRIVATE */ + +extern vm_map_t kernel_map; +extern vm_map_t kernel_pageable_map; extern vm_map_t ipc_kernel_map; -#endif /* KERNEL_PRIVATE */ +#endif /* KERNEL_PRIVATE */ + +#ifdef KERNEL + +__BEGIN_DECLS +#if MACH_KERNEL_PRIVATE +extern vm_offset_t vm_kernel_addrhash(vm_offset_t addr) +__XNU_INTERNAL(vm_kernel_addrhash); +#else +extern vm_offset_t vm_kernel_addrhash(vm_offset_t addr); +#endif +__END_DECLS + +extern void vm_kernel_addrhide( + vm_offset_t addr, + vm_offset_t *hide_addr); + +extern vm_offset_t vm_kernel_addrperm_ext; + +extern void vm_kernel_addrperm_external( + vm_offset_t addr, + vm_offset_t *perm_addr); + +extern void vm_kernel_unslide_or_perm_external( + vm_offset_t addr, + vm_offset_t *up_addr); + +#if MACH_KERNEL_PRIVATE +extern uint64_t vm_kernel_addrhash_salt; +extern uint64_t vm_kernel_addrhash_salt_ext; + +extern void vm_kernel_addrhash_external( + vm_offset_t addr, + vm_offset_t *perm_addr); +#endif /* MACH_KERNEL_PRIVATE */ + +extern void vm_init_before_launchd(void); + +#endif /* KERNEL */ + +#ifdef __cplusplus +} +#endif -#endif /* _VM_VM_KERN_H_ */ +#endif /* _VM_VM_KERN_H_ */