X-Git-Url: https://git.saurik.com/apple/xnu.git/blobdiff_plain/1c79356b52d46aa6b508fb032f5ae709b1f2897b..bd504ef0e0b883cdd7917b73b3574eb9ce669905:/osfmk/mach/i386/vm_param.h?ds=sidebyside diff --git a/osfmk/mach/i386/vm_param.h b/osfmk/mach/i386/vm_param.h index f5b0c0536..12eb226e2 100644 --- a/osfmk/mach/i386/vm_param.h +++ b/osfmk/mach/i386/vm_param.h @@ -1,23 +1,29 @@ /* - * Copyright (c) 2000 Apple Computer, Inc. All rights reserved. + * Copyright (c) 2000-2006 Apple Computer, Inc. All rights reserved. * - * @APPLE_LICENSE_HEADER_START@ + * @APPLE_OSREFERENCE_LICENSE_HEADER_START@ * - * The contents of this file constitute Original Code as defined in and - * are subject to the Apple Public Source License Version 1.1 (the - * "License"). You may not use this file except in compliance with the - * License. Please obtain a copy of the License at - * http://www.apple.com/publicsource and read it before using this file. + * This file contains Original Code and/or Modifications of Original Code + * as defined in and that are subject to the Apple Public Source License + * Version 2.0 (the 'License'). You may not use this file except in + * compliance with the License. The rights granted to you under the License + * may not be used to create, or enable the creation or redistribution of, + * unlawful or unlicensed copies of an Apple operating system, or to + * circumvent, violate, or enable the circumvention or violation of, any + * terms of an Apple operating system software license agreement. * - * This Original Code and all software distributed under the License are - * distributed on an "AS IS" basis, WITHOUT WARRANTY OF ANY KIND, EITHER + * Please obtain a copy of the License at + * http://www.opensource.apple.com/apsl/ and read it before using this file. + * + * The Original Code and all software distributed under the License are + * distributed on an 'AS IS' basis, WITHOUT WARRANTY OF ANY KIND, EITHER * EXPRESS OR IMPLIED, AND APPLE HEREBY DISCLAIMS ALL SUCH WARRANTIES, * INCLUDING WITHOUT LIMITATION, ANY WARRANTIES OF MERCHANTABILITY, - * FITNESS FOR A PARTICULAR PURPOSE OR NON-INFRINGEMENT. Please see the - * License for the specific language governing rights and limitations - * under the License. + * FITNESS FOR A PARTICULAR PURPOSE, QUIET ENJOYMENT OR NON-INFRINGEMENT. + * Please see the License for the specific language governing rights and + * limitations under the License. * - * @APPLE_LICENSE_HEADER_END@ + * @APPLE_OSREFERENCE_LICENSE_HEADER_END@ */ /* * @OSF_COPYRIGHT@ @@ -84,19 +90,28 @@ #ifndef _MACH_I386_VM_PARAM_H_ #define _MACH_I386_VM_PARAM_H_ -#define BYTE_SIZE 8 /* byte size in bits */ +#define BYTE_SIZE 8 /* byte size in bits */ + +#define I386_PGBYTES 4096 /* bytes per 80386 page */ +#define I386_PGSHIFT 12 /* bitshift for pages */ -#define I386_PGBYTES 4096 /* bytes per 80386 page */ -#define I386_PGSHIFT 12 /* number of bits to shift for pages */ +#define PAGE_SIZE I386_PGBYTES +#define PAGE_SHIFT I386_PGSHIFT +#define PAGE_MASK (PAGE_SIZE - 1) + +#define I386_LPGBYTES 2*1024*1024 /* bytes per large page */ +#define I386_LPGSHIFT 21 /* bitshift for large pages */ +#define I386_LPGMASK (I386_LPGBYTES-1) /* * Convert bytes to pages and convert pages to bytes. * No rounding is used. */ -#define i386_btop(x) (((unsigned)(x)) >> I386_PGSHIFT) +#define i386_btop(x) ((ppnum_t)((x) >> I386_PGSHIFT)) #define machine_btop(x) i386_btop(x) -#define i386_ptob(x) (((unsigned)(x)) << I386_PGSHIFT) +#define i386_ptob(x) (((pmap_paddr_t)(x)) << I386_PGSHIFT) +#define machine_ptob(x) i386_ptob(x) /* * Round off or truncate to the nearest page. These will work @@ -104,48 +119,151 @@ * bytes. */ -#define i386_round_page(x) ((((unsigned)(x)) + I386_PGBYTES - 1) & \ +#define i386_round_page(x) ((((pmap_paddr_t)(x)) + I386_PGBYTES - 1) & \ ~(I386_PGBYTES-1)) -#define i386_trunc_page(x) (((unsigned)(x)) & ~(I386_PGBYTES-1)) +#define i386_trunc_page(x) (((pmap_paddr_t)(x)) & ~(I386_PGBYTES-1)) + + +#define VM_MIN_ADDRESS64 ((user_addr_t) 0x0000000000000000ULL) +/* + * default top of user stack... it grows down from here + */ +#define VM_USRSTACK64 ((user_addr_t) 0x00007FFF5FC00000ULL) +#define VM_DYLD64 ((user_addr_t) 0x00007FFF5FC00000ULL) +#define VM_LIB64_SHR_DATA ((user_addr_t) 0x00007FFF60000000ULL) +#define VM_LIB64_SHR_TEXT ((user_addr_t) 0x00007FFF80000000ULL) +/* + * the end of the usable user address space , for now about 47 bits. + * the 64 bit commpage is past the end of this + */ +#define VM_MAX_PAGE_ADDRESS ((user_addr_t) 0x00007FFFFFE00000ULL) +/* + * canonical end of user address space for limits checking + */ +#define VM_MAX_USER_PAGE_ADDRESS ((user_addr_t)0x00007FFFFFFFF000ULL) + + +/* system-wide values */ +#define MACH_VM_MIN_ADDRESS ((mach_vm_offset_t) 0) +#define MACH_VM_MAX_ADDRESS ((mach_vm_offset_t) VM_MAX_PAGE_ADDRESS) + +/* process-relative values (all 32-bit legacy only for now) */ #define VM_MIN_ADDRESS ((vm_offset_t) 0) -#define VM_MAX_ADDRESS ((vm_offset_t) 0xc0000000U) +#define VM_USRSTACK32 ((vm_offset_t) 0xC0000000) /* ASLR slides stack down by up to 1 MB */ +#define VM_MAX_ADDRESS ((vm_offset_t) 0xFFE00000) + + +#ifdef KERNEL_PRIVATE + +/* Kernel-wide values */ + +#define KB (1024ULL) +#define MB (1024*KB) +#define GB (1024*MB) + +/* + * Maximum physical memory supported. + */ +#define K32_MAXMEM (32*GB) +#define K64_MAXMEM (96*GB) +#if defined(__i386__) +#define KERNEL_MAXMEM K32_MAXMEM +#else +#define KERNEL_MAXMEM K64_MAXMEM +#endif + +/* + * XXX + * The kernel max VM address is limited to 0xFF3FFFFF for now because + * some data structures are explicitly allocated at 0xFF400000 without + * VM's knowledge (see osfmk/i386/locore.s for the allocation of PTmap and co.). + * We can't let VM allocate memory from there. + */ + +#if defined(__i386__) + +#define KERNEL_IMAGE_TO_PHYS(x) (x) +#define VM_MIN_KERNEL_ADDRESS ((vm_offset_t) 0x00001000U) +#define VM_MIN_KERNEL_AND_KEXT_ADDRESS VM_MIN_KERNEL_ADDRESS +#define VM_MAX_KERNEL_ADDRESS ((vm_offset_t) 0xFE7FFFFFU) + +#elif defined(__x86_64__) + +#define KERNEL_IMAGE_TO_PHYS(x) (x) +#define VM_MIN_KERNEL_ADDRESS ((vm_offset_t) 0xFFFFFF8000000000UL) +#define VM_MIN_KERNEL_PAGE ((ppnum_t)0) +#define VM_MIN_KERNEL_AND_KEXT_ADDRESS (VM_MIN_KERNEL_ADDRESS - 0x80000000ULL) +#define VM_MAX_KERNEL_ADDRESS ((vm_offset_t) 0xFFFFFFFFFFFFEFFFUL) +#define VM_MAX_KERNEL_ADDRESS_EFI32 ((vm_offset_t) 0xFFFFFF80FFFFEFFFUL) +#define KEXT_ALLOC_MAX_OFFSET (2 * 1024 * 1024 * 1024UL) +#define KEXT_ALLOC_BASE(x) ((x) - KEXT_ALLOC_MAX_OFFSET) +#define KEXT_ALLOC_SIZE(x) (KEXT_ALLOC_MAX_OFFSET - (x)) -#define LINEAR_KERNEL_ADDRESS ((vm_offset_t) 0xc0000000) +#define VM_KERNEL_IS_KEXT(_o) \ + (((vm_offset_t)(_o) >= VM_MIN_KERNEL_AND_KEXT_ADDRESS) && \ + ((vm_offset_t)(_o) < VM_MIN_KERNEL_ADDRESS)) -#define VM_MIN_KERNEL_ADDRESS ((vm_offset_t) 0x00000000U) -#define VM_MAX_KERNEL_ADDRESS ((vm_offset_t) 0x3fffffffU) +#else +#error unsupported architecture +#endif -#define VM_MIN_KERNEL_LOADED_ADDRESS ((vm_offset_t) 0x0c000000U) -#define VM_MAX_KERNEL_LOADED_ADDRESS ((vm_offset_t) 0x1fffffffU) +#define KERNEL_STACK_SIZE (I386_PGBYTES*4) + +#define VM_MAP_MIN_ADDRESS MACH_VM_MIN_ADDRESS +#define VM_MAP_MAX_ADDRESS MACH_VM_MAX_ADDRESS /* FIXME - always leave like this? */ #define INTSTACK_SIZE (I386_PGBYTES*4) -#define KERNEL_STACK_SIZE (I386_PGBYTES*4) -#if 0 /* FIXME */ +#ifdef MACH_KERNEL_PRIVATE -#include -#include -#include +/* For implementing legacy 32-bit interfaces */ +#define VM32_SUPPORT 1 +#define VM32_MIN_ADDRESS ((vm32_offset_t) 0) +#define VM32_MAX_ADDRESS ((vm32_offset_t) (VM_MAX_PAGE_ADDRESS & 0xFFFFFFFF)) + +/* + * kalloc() parameters: + * + * Historically kalloc's underlying zones were power-of-2 sizes, with a + * KALLOC_MINSIZE of 16 bytes. The allocator ensured that + * (sizeof == alignof) >= 16 for all kalloc allocations. + * + * Today kalloc may use zones with intermediate sizes, constrained by + * KALLOC_MINSIZE and a minimum alignment, expressed by KALLOC_LOG2_MINALIGN. + * + * The common alignment for LP64 is for longs and pointers i.e. 8 bytes. + */ + +#if defined(__i386__) + +#define KALLOC_MINSIZE 16 /* minimum allocation size */ +#define KALLOC_LOG2_MINALIGN 4 /* log2 minimum alignment */ + +#define LINEAR_KERNEL_ADDRESS ((vm_offset_t) 0x00000000) + +#define VM_MIN_KERNEL_LOADED_ADDRESS ((vm_offset_t) 0x00000000U) +#define VM_MAX_KERNEL_LOADED_ADDRESS ((vm_offset_t) 0x1FFFFFFFU) + +#define NCOPY_WINDOWS 4 + +#elif defined(__x86_64__) + +#define KALLOC_MINSIZE 16 /* minimum allocation size */ +#define KALLOC_LOG2_MINALIGN 4 /* log2 minimum alignment */ + +#define LINEAR_KERNEL_ADDRESS ((vm_offset_t) 0x00000000) + +#define VM_MIN_KERNEL_LOADED_ADDRESS ((vm_offset_t) 0xFFFFFF8000000000UL) +#define VM_MAX_KERNEL_LOADED_ADDRESS ((vm_offset_t) 0xFFFFFF801FFFFFFFUL) + +#define NCOPY_WINDOWS 0 -#if defined(AT386) -#include -#endif -#if !NORMA_VM -#if !TASK_SWAPPER && !THREAD_SWAPPER -#define KERNEL_STACK_SIZE (I386_PGBYTES/2) #else -/* stack needs to be a multiple of page size to get unwired when swapped */ -#define KERNEL_STACK_SIZE (I386_PGBYTES) -#endif /* TASK || THREAD SWAPPER */ -#define INTSTACK_SIZE (I386_PGBYTES) /* interrupt stack size */ -#else /* NORMA_VM */ -#define KERNEL_STACK_SIZE (I386_PGBYTES*2) -#define INTSTACK_SIZE (I386_PGBYTES*2) /* interrupt stack size */ -#endif /* NORMA_VM */ -#endif /* MACH_KERNEL */ +#error unsupported architecture +#endif /* * Conversion between 80386 pages and VM pages @@ -155,54 +273,26 @@ #define round_i386_to_vm(p) (atop(round_page(i386_ptob(p)))) #define vm_to_i386(p) (i386_btop(ptoa(p))) -/* - * Physical memory is mapped 1-1 with virtual memory starting - * at VM_MIN_KERNEL_ADDRESS. - */ -#define phystokv(a) ((vm_offset_t)(a) + VM_MIN_KERNEL_ADDRESS) -/* - * For 386 only, ensure that pages are installed in the - * kernel_pmap with VM_PROT_WRITE enabled. This avoids - * code in pmap_enter that disallows a read-only mapping - * in the kernel's pmap. (See ri-osc CR1387.) - * - * An entry in kernel_pmap is made only by the kernel or - * a collocated server -- by definition (;-)), the requester - * is trusted code. If it asked for read-only access, - * it won't attempt a write. We don't have to enforce the - * restriction. (Naturally, this assumes that any collocated - * server will _not_ depend on trapping write accesses to pages - * mapped read-only; this cannot be made to work in the current - * i386-inspired pmap model.) - */ - -/*#if defined(AT386) +#define PMAP_SET_CACHE_ATTR(mem, object, cache_attr, batch_pmap_op) \ + MACRO_BEGIN \ + pmap_set_cache_attributes((mem)->phys_page, (cache_attr)); \ + (object)->set_cache_attr = TRUE; \ + (void) batch_pmap_op; \ + MACRO_END -#define PMAP_ENTER_386_CHECK \ - if (cpuid_family == CPUID_FAMILY_386) - -#else -- FIXME? We're only running on Pentiums or better */ +#define PMAP_BATCH_SET_CACHE_ATTR(object, user_page_list, cache_attr, num_pages, batch_pmap_op)\ + MACRO_BEGIN \ + (void) user_page_list; \ + (void) num_pages; \ + (void) batch_pmap_op; \ + MACRO_END -#define PMAP_ENTER_386_CHECK +#define IS_USERADDR64_CANONICAL(addr) \ + ((addr) < (VM_MAX_USER_PAGE_ADDRESS + PAGE_SIZE)) -/*#endif*/ +#endif /* MACH_KERNEL_PRIVATE */ -#define PMAP_ENTER(pmap, virtual_address, page, protection, wired) \ - MACRO_BEGIN \ - vm_prot_t __prot__ = \ - (protection) & ~(page)->page_lock; \ - \ - PMAP_ENTER_386_CHECK \ - if ((pmap) == kernel_pmap) \ - __prot__ |= VM_PROT_WRITE; \ - pmap_enter( \ - (pmap), \ - (virtual_address), \ - (page)->phys_addr, \ - __prot__, \ - (wired) \ - ); \ - MACRO_END +#endif /* KERNEL_PRIVATE */ #endif /* _MACH_I386_VM_PARAM_H_ */