X-Git-Url: https://git.saurik.com/apple/xnu.git/blobdiff_plain/de355530ae67247cbd0da700edb3a2a1dae884c2..4a3eedf9ecc9bbe3f3a5c6ce5e53ad199d639d32:/osfmk/ppc/io_map.c?ds=sidebyside diff --git a/osfmk/ppc/io_map.c b/osfmk/ppc/io_map.c index 5ce5ab5e8..25607c1cc 100644 --- a/osfmk/ppc/io_map.c +++ b/osfmk/ppc/io_map.c @@ -1,23 +1,29 @@ /* * Copyright (c) 2000 Apple Computer, Inc. All rights reserved. * - * @APPLE_LICENSE_HEADER_START@ + * @APPLE_OSREFERENCE_LICENSE_HEADER_START@ * - * The contents of this file constitute Original Code as defined in and - * are subject to the Apple Public Source License Version 1.1 (the - * "License"). You may not use this file except in compliance with the - * License. Please obtain a copy of the License at - * http://www.apple.com/publicsource and read it before using this file. + * This file contains Original Code and/or Modifications of Original Code + * as defined in and that are subject to the Apple Public Source License + * Version 2.0 (the 'License'). You may not use this file except in + * compliance with the License. The rights granted to you under the License + * may not be used to create, or enable the creation or redistribution of, + * unlawful or unlicensed copies of an Apple operating system, or to + * circumvent, violate, or enable the circumvention or violation of, any + * terms of an Apple operating system software license agreement. * - * This Original Code and all software distributed under the License are - * distributed on an "AS IS" basis, WITHOUT WARRANTY OF ANY KIND, EITHER + * Please obtain a copy of the License at + * http://www.opensource.apple.com/apsl/ and read it before using this file. + * + * The Original Code and all software distributed under the License are + * distributed on an 'AS IS' basis, WITHOUT WARRANTY OF ANY KIND, EITHER * EXPRESS OR IMPLIED, AND APPLE HEREBY DISCLAIMS ALL SUCH WARRANTIES, * INCLUDING WITHOUT LIMITATION, ANY WARRANTIES OF MERCHANTABILITY, - * FITNESS FOR A PARTICULAR PURPOSE OR NON-INFRINGEMENT. Please see the - * License for the specific language governing rights and limitations - * under the License. + * FITNESS FOR A PARTICULAR PURPOSE, QUIET ENJOYMENT OR NON-INFRINGEMENT. + * Please see the License for the specific language governing rights and + * limitations under the License. * - * @APPLE_LICENSE_HEADER_END@ + * @APPLE_OSREFERENCE_LICENSE_HEADER_END@ */ /* * @OSF_COPYRIGHT@ @@ -42,54 +48,84 @@ extern vm_offset_t virtual_avail; * outside the usual physical memory. If phys_addr is NULL then * steal the appropriate number of physical pages from the vm * system and map them. + * + * Note, this will onl */ vm_offset_t -io_map(phys_addr, size) - vm_offset_t phys_addr; - vm_size_t size; +io_map(vm_offset_t phys_addr, vm_size_t size, unsigned int flags) { vm_offset_t start; - int i; - unsigned int j; + vm_size_t i; + unsigned int mflags; vm_page_t m; + mflags = mmFlgBlock | mmFlgUseAttr | (flags & VM_MEM_GUARDED) | ((flags & VM_MEM_NOT_CACHEABLE) >> 1); /* Convert to our mapping_make flags */ #if DEBUG assert (kernel_map != VM_MAP_NULL); /* VM must be initialised */ #endif - if (phys_addr != 0) { - /* make sure we map full contents of all the pages concerned */ - size = round_page(size + (phys_addr & PAGE_MASK)); + if (phys_addr != 0) { /* If they supplied a physical address, use it */ - /* Steal some free virtual addresses */ + size = round_page(size + (phys_addr & PAGE_MASK)); /* Make sure we map all of it */ - (void) kmem_alloc_pageable(kernel_map, &start, size); + (void) kmem_alloc_pageable(kernel_map, &start, size); /* Get some virtual addresses to use */ - pmap_map_block(kernel_pmap, start, phys_addr, size, - VM_PROT_READ|VM_PROT_WRITE, PTE_WIMG_IO, 0); /* Set up a block mapped area */ + (void)mapping_make(kernel_pmap, (addr64_t)start, (ppnum_t)(phys_addr >> 12), + mflags, /* Map with requested cache mode */ + (size >> 12), VM_PROT_READ|VM_PROT_WRITE); - return (start + (phys_addr & PAGE_MASK)); + return (start + (phys_addr & PAGE_MASK)); /* Pass back the physical address */ } else { - /* Steal some free virtual addresses */ - (void) kmem_alloc_pageable(kernel_map, &start, size); + (void) kmem_alloc_pageable(kernel_map, &start, size); /* Get some virtual addresses */ mapping_prealloc(size); /* Make sure there are enough free mappings */ - /* Steal some physical pages and map them one by one */ + for (i = 0; i < size ; i += PAGE_SIZE) { m = VM_PAGE_NULL; - while ((m = vm_page_grab()) == VM_PAGE_NULL) - VM_PAGE_WAIT(); + while ((m = vm_page_grab()) == VM_PAGE_NULL) { /* Get a physical page */ + VM_PAGE_WAIT(); /* Wait if we didn't have one */ + } vm_page_gobble(m); - (void) pmap_map_bd(start + i, - m->phys_addr, - m->phys_addr + PAGE_SIZE, - VM_PROT_READ|VM_PROT_WRITE); + + (void)mapping_make(kernel_pmap, + (addr64_t)(start + i), m->phys_page, + mflags, /* Map with requested cache mode */ + 1, VM_PROT_READ|VM_PROT_WRITE); + } mapping_relpre(); /* Allow mapping release */ return start; } } + + +/* + * Allocate and map memory for devices before the VM system comes alive. + */ + +vm_offset_t io_map_spec(vm_offset_t phys_addr, vm_size_t size, unsigned int flags) +{ + vm_offset_t start; + unsigned int mflags; + + if(kernel_map != VM_MAP_NULL) { /* If VM system is up, redirect to normal routine */ + + return io_map(phys_addr, size, flags); /* Map the address */ + + } + + mflags = mmFlgBlock | mmFlgUseAttr | (flags & VM_MEM_GUARDED) | ((flags & VM_MEM_NOT_CACHEABLE) >> 1); /* Convert to our mapping_make flags */ + + size = round_page(size + (phys_addr - (phys_addr & -PAGE_SIZE))); /* Extend the length to include it all */ + start = pmap_boot_map(size); /* Get me some virtual address */ + + (void)mapping_make(kernel_pmap, (addr64_t)start, (ppnum_t)(phys_addr >> 12), + mflags, /* Map with requested cache mode */ + (size >> 12), VM_PROT_READ|VM_PROT_WRITE); + + return (start + (phys_addr & PAGE_MASK)); +}