X-Git-Url: https://git.saurik.com/apple/xnu.git/blobdiff_plain/4452a7af2eac33dbad800bcc91f2399d62c18f53..39236c6e673c41db228275375ab7fdb0f837b292:/osfmk/mach/i386/vm_param.h?ds=sidebyside diff --git a/osfmk/mach/i386/vm_param.h b/osfmk/mach/i386/vm_param.h index b7d6931ab..02bef271f 100644 --- a/osfmk/mach/i386/vm_param.h +++ b/osfmk/mach/i386/vm_param.h @@ -1,5 +1,5 @@ /* - * Copyright (c) 2000-2004 Apple Computer, Inc. All rights reserved. + * Copyright (c) 2000-2012 Apple Inc. All rights reserved. * * @APPLE_OSREFERENCE_LICENSE_HEADER_START@ * @@ -92,21 +92,32 @@ #define BYTE_SIZE 8 /* byte size in bits */ -#define I386_PGBYTES 4096 /* bytes per 80386 page */ -#define I386_PGSHIFT 12 /* number of bits to shift for pages */ +#define I386_PGBYTES 4096 /* bytes per 80386 page */ +#define I386_PGSHIFT 12 /* bitshift for pages */ + +#ifdef PRIVATE +#define KERNEL_PAGE_SIZE I386_PGBYTES +#define KERNEL_PAGE_SHIFT I386_PGSHIFT +#define KERNEL_PAGE_MASK (KERNEL_PAGE_SIZE-1) +#endif #define PAGE_SIZE I386_PGBYTES #define PAGE_SHIFT I386_PGSHIFT #define PAGE_MASK (PAGE_SIZE - 1) +#define I386_LPGBYTES 2*1024*1024 /* bytes per large page */ +#define I386_LPGSHIFT 21 /* bitshift for large pages */ +#define I386_LPGMASK (I386_LPGBYTES-1) + /* * Convert bytes to pages and convert pages to bytes. * No rounding is used. */ -#define i386_btop(x) (((pmap_paddr_t)(x)) >> I386_PGSHIFT) +#define i386_btop(x) ((ppnum_t)((x) >> I386_PGSHIFT)) #define machine_btop(x) i386_btop(x) #define i386_ptob(x) (((pmap_paddr_t)(x)) << I386_PGSHIFT) +#define machine_ptob(x) i386_ptob(x) /* * Round off or truncate to the nearest page. These will work @@ -139,23 +150,31 @@ #define VM_MAX_USER_PAGE_ADDRESS ((user_addr_t)0x00007FFFFFFFF000ULL) - /* system-wide values */ #define MACH_VM_MIN_ADDRESS ((mach_vm_offset_t) 0) #define MACH_VM_MAX_ADDRESS ((mach_vm_offset_t) VM_MAX_PAGE_ADDRESS) /* process-relative values (all 32-bit legacy only for now) */ #define VM_MIN_ADDRESS ((vm_offset_t) 0) -#define VM_USRSTACK32 ((vm_offset_t) 0xC0000000) +#define VM_USRSTACK32 ((vm_offset_t) 0xC0000000) /* ASLR slides stack down by up to 1 MB */ #define VM_MAX_ADDRESS ((vm_offset_t) 0xFFE00000) - - #ifdef KERNEL_PRIVATE /* Kernel-wide values */ -#define VM_MIN_KERNEL_ADDRESS ((vm_offset_t) 0x00001000U) + +#define KB (1024ULL) +#define MB (1024*KB) +#define GB (1024*MB) + +/* + * Maximum physical memory supported. + */ +#define K32_MAXMEM (32*GB) +#define K64_MAXMEM (128*GB) +#define KERNEL_MAXMEM K64_MAXMEM + /* * XXX * The kernel max VM address is limited to 0xFF3FFFFF for now because @@ -164,8 +183,19 @@ * We can't let VM allocate memory from there. */ -#define VM_MAX_KERNEL_ADDRESS ((vm_offset_t) 0xFE7FFFFF) -#define KERNEL_STACK_SIZE (I386_PGBYTES*4) + +#define KERNEL_IMAGE_TO_PHYS(x) (x) +#define VM_MIN_KERNEL_ADDRESS ((vm_offset_t) 0xFFFFFF8000000000UL) +#define VM_MIN_KERNEL_PAGE ((ppnum_t)0) +#define VM_MIN_KERNEL_AND_KEXT_ADDRESS (VM_MIN_KERNEL_ADDRESS - 0x80000000ULL) +#define VM_MAX_KERNEL_ADDRESS ((vm_offset_t) 0xFFFFFFFFFFFFEFFFUL) +#define VM_MAX_KERNEL_ADDRESS_EFI32 ((vm_offset_t) 0xFFFFFF80FFFFEFFFUL) +#define KEXT_ALLOC_MAX_OFFSET (2 * 1024 * 1024 * 1024UL) +#define KEXT_ALLOC_BASE(x) ((x) - KEXT_ALLOC_MAX_OFFSET) +#define KEXT_ALLOC_SIZE(x) (KEXT_ALLOC_MAX_OFFSET - (x)) + + +#define KERNEL_STACK_SIZE (I386_PGBYTES*4) #define VM_MAP_MIN_ADDRESS MACH_VM_MIN_ADDRESS #define VM_MAP_MAX_ADDRESS MACH_VM_MAX_ADDRESS @@ -176,16 +206,35 @@ #ifdef MACH_KERNEL_PRIVATE /* For implementing legacy 32-bit interfaces */ -#define VM32_SUPPORT +#define VM32_SUPPORT 1 #define VM32_MIN_ADDRESS ((vm32_offset_t) 0) #define VM32_MAX_ADDRESS ((vm32_offset_t) (VM_MAX_PAGE_ADDRESS & 0xFFFFFFFF)) +/* + * kalloc() parameters: + * + * Historically kalloc's underlying zones were power-of-2 sizes, with a + * KALLOC_MINSIZE of 16 bytes. The allocator ensured that + * (sizeof == alignof) >= 16 for all kalloc allocations. + * + * Today kalloc may use zones with intermediate sizes, constrained by + * KALLOC_MINSIZE and a minimum alignment, expressed by KALLOC_LOG2_MINALIGN. + * + * The common alignment for LP64 is for longs and pointers i.e. 8 bytes. + */ + + +#define KALLOC_MINSIZE 16 /* minimum allocation size */ +#define KALLOC_LOG2_MINALIGN 4 /* log2 minimum alignment */ + #define LINEAR_KERNEL_ADDRESS ((vm_offset_t) 0x00000000) -#define VM_MIN_KERNEL_LOADED_ADDRESS ((vm_offset_t) 0x00000000U) -#define VM_MAX_KERNEL_LOADED_ADDRESS ((vm_offset_t) 0x1fffffffU) +#define VM_MIN_KERNEL_LOADED_ADDRESS ((vm_offset_t) 0xFFFFFF8000000000UL) +#define VM_MAX_KERNEL_LOADED_ADDRESS ((vm_offset_t) 0xFFFFFF801FFFFFFFUL) + +#define NCOPY_WINDOWS 0 + -#define NCOPY_WINDOWS 4 /* * Conversion between 80386 pages and VM pages @@ -195,27 +244,19 @@ #define round_i386_to_vm(p) (atop(round_page(i386_ptob(p)))) #define vm_to_i386(p) (i386_btop(ptoa(p))) -#define PMAP_ENTER(pmap, virtual_address, page, protection, flags, wired) \ - MACRO_BEGIN \ - pmap_t __pmap = (pmap); \ - vm_page_t __page = (page); \ - vm_prot_t __prot__ = \ - (protection) & ~(page)->page_lock; \ - \ - if (__pmap == kernel_pmap) { \ - __prot__ |= VM_PROT_WRITE; \ - } else { \ - assert(!__page->encrypted); \ - } \ - \ - pmap_enter( \ - __pmap, \ - (virtual_address), \ - __page->phys_page, \ - __prot__, \ - flags, \ - (wired) \ - ); \ + +#define PMAP_SET_CACHE_ATTR(mem, object, cache_attr, batch_pmap_op) \ + MACRO_BEGIN \ + pmap_set_cache_attributes((mem)->phys_page, (cache_attr)); \ + (object)->set_cache_attr = TRUE; \ + (void) batch_pmap_op; \ + MACRO_END + +#define PMAP_BATCH_SET_CACHE_ATTR(object, user_page_list, cache_attr, num_pages, batch_pmap_op)\ + MACRO_BEGIN \ + (void) user_page_list; \ + (void) num_pages; \ + (void) batch_pmap_op; \ MACRO_END #define IS_USERADDR64_CANONICAL(addr) \