X-Git-Url: https://git.saurik.com/apple/xnu.git/blobdiff_plain/c0fea4742e91338fffdcf79f86a7c1d5e2b97eb1..6601e61aa18bf4f09af135ff61fc7f4771d23b06:/osfmk/kern/kalloc.c?ds=sidebyside diff --git a/osfmk/kern/kalloc.c b/osfmk/kern/kalloc.c index 3d1d0c899..57161e0bb 100644 --- a/osfmk/kern/kalloc.c +++ b/osfmk/kern/kalloc.c @@ -80,7 +80,6 @@ vm_map_t kalloc_map; vm_size_t kalloc_map_size = 16 * 1024 * 1024; vm_size_t kalloc_max; vm_size_t kalloc_max_prerounded; -vm_size_t kalloc_kernmap_size; /* size of kallocs that can come from kernel map */ unsigned int kalloc_large_inuse; vm_size_t kalloc_large_total; @@ -99,8 +98,6 @@ vm_size_t kalloc_large_max; * It represents the first power of two for which no zone exists. * kalloc_max_prerounded is the smallest allocation size, before * rounding, for which no zone exists. - * Also if the allocation size is more than kalloc_kernmap_size - * then allocate from kernel map rather than kalloc_map. */ int first_k_zone = -1; @@ -192,8 +189,6 @@ kalloc_init( else kalloc_max = PAGE_SIZE; kalloc_max_prerounded = kalloc_max / 2 + 1; - /* size it to be more than 16 times kalloc_max (256k) for allocations from kernel map */ - kalloc_kernmap_size = (kalloc_max * 16) + 1; /* * Allocate a zone for each size we are going to handle. @@ -220,7 +215,6 @@ kalloc_canblock( { register int zindex; register vm_size_t allocsize; - vm_map_t alloc_map = VM_MAP_NULL; /* * If size is too large for a zone, then use kmem_alloc. @@ -235,13 +229,7 @@ kalloc_canblock( if (!canblock) { return(0); } - - if (size >= kalloc_kernmap_size) - alloc_map = kernel_map; - else - alloc_map = kalloc_map; - - if (kmem_alloc(alloc_map, (vm_offset_t *)&addr, size) != KERN_SUCCESS) + if (kmem_alloc(kalloc_map, (vm_offset_t *)&addr, size) != KERN_SUCCESS) addr = 0; if (addr) { @@ -293,7 +281,6 @@ krealloc( register int zindex; register vm_size_t allocsize; void *naddr; - vm_map_t alloc_map = VM_MAP_NULL; /* can only be used for increasing allocation size */ @@ -312,16 +299,11 @@ krealloc( /* if old block was kmem_alloc'd, then use kmem_realloc if necessary */ if (old_size >= kalloc_max_prerounded) { - if (old_size >= kalloc_kernmap_size) - alloc_map = kernel_map; - else - alloc_map = kalloc_map; - old_size = round_page(old_size); new_size = round_page(new_size); if (new_size > old_size) { - if (KERN_SUCCESS != kmem_realloc(alloc_map, + if (KERN_SUCCESS != kmem_realloc(kalloc_map, (vm_offset_t)*addrp, old_size, (vm_offset_t *)&naddr, new_size)) { panic("krealloc: kmem_realloc"); @@ -332,7 +314,7 @@ krealloc( *addrp = (void *) naddr; /* kmem_realloc() doesn't free old page range. */ - kmem_free(alloc_map, (vm_offset_t)*addrp, old_size); + kmem_free(kalloc_map, (vm_offset_t)*addrp, old_size); kalloc_large_total += (new_size - old_size); @@ -362,11 +344,7 @@ krealloc( simple_unlock(lock); if (new_size >= kalloc_max_prerounded) { - if (new_size >= kalloc_kernmap_size) - alloc_map = kernel_map; - else - alloc_map = kalloc_map; - if (KERN_SUCCESS != kmem_alloc(alloc_map, + if (KERN_SUCCESS != kmem_alloc(kalloc_map, (vm_offset_t *)&naddr, new_size)) { panic("krealloc: kmem_alloc"); simple_lock(lock); @@ -441,16 +419,11 @@ kfree( { register int zindex; register vm_size_t freesize; - vm_map_t alloc_map = VM_MAP_NULL; /* if size was too large for a zone, then use kmem_free */ if (size >= kalloc_max_prerounded) { - if (size >= kalloc_kernmap_size) - alloc_map = kernel_map; - else - alloc_map = kalloc_map; - kmem_free(alloc_map, (vm_offset_t)data, size); + kmem_free(kalloc_map, (vm_offset_t)data, size); kalloc_large_total -= size; kalloc_large_inuse--;