- /* Allocate some kernel address space. */
- _buffer = IOMallocPageable(capacity, alignment);
- /* Allocate a wired-down buffer inside kernel space. */
- else if (options & kIOMemoryPhysicallyContiguous)
- _buffer = IOMallocContiguous(capacity, alignment, 0);
- else if (alignment > 1)
- _buffer = IOMallocAligned(capacity, alignment);
- else
- _buffer = IOMalloc(capacity);
+ {
+ if (inTask == kernel_task)
+ {
+ /* Allocate some kernel address space. */
+ _buffer = IOMallocPageable(capacity, alignment);
+ if (_buffer)
+ map = IOPageableMapForAddress((vm_address_t) _buffer);
+ }
+ else
+ {
+ kern_return_t kr;
+
+ if( !reserved) {
+ reserved = IONew( ExpansionData, 1 );
+ if( !reserved)
+ return( false );
+ }
+ map = get_task_map(inTask);
+ vm_map_reference(map);
+ reserved->map = map;
+ kr = vm_allocate( map, (vm_address_t *) &_buffer, round_page(capacity),
+ VM_FLAGS_ANYWHERE | VM_MAKE_TAG(VM_MEMORY_IOKIT) );
+ if( KERN_SUCCESS != kr)
+ return( false );
+
+ // we have to make sure that these pages don't get copied on fork.
+ kr = vm_inherit( map, (vm_address_t) _buffer, round_page(capacity), VM_INHERIT_NONE);
+ if( KERN_SUCCESS != kr)
+ return( false );
+ }
+ }
+ else
+ {
+ /* Allocate a wired-down buffer inside kernel space. */
+ if (options & kIOMemoryPhysicallyContiguous)
+ _buffer = IOMallocContiguous(capacity, alignment, 0);
+ else if (alignment > 1)
+ _buffer = IOMallocAligned(capacity, alignment);
+ else
+ _buffer = IOMalloc(capacity);
+ }