]>
Commit | Line | Data |
---|---|---|
1 | /* | |
2 | * Copyright (c) 2000 Apple Computer, Inc. All rights reserved. | |
3 | * | |
4 | * @APPLE_OSREFERENCE_LICENSE_HEADER_START@ | |
5 | * | |
6 | * This file contains Original Code and/or Modifications of Original Code | |
7 | * as defined in and that are subject to the Apple Public Source License | |
8 | * Version 2.0 (the 'License'). You may not use this file except in | |
9 | * compliance with the License. The rights granted to you under the License | |
10 | * may not be used to create, or enable the creation or redistribution of, | |
11 | * unlawful or unlicensed copies of an Apple operating system, or to | |
12 | * circumvent, violate, or enable the circumvention or violation of, any | |
13 | * terms of an Apple operating system software license agreement. | |
14 | * | |
15 | * Please obtain a copy of the License at | |
16 | * http://www.opensource.apple.com/apsl/ and read it before using this file. | |
17 | * | |
18 | * The Original Code and all software distributed under the License are | |
19 | * distributed on an 'AS IS' basis, WITHOUT WARRANTY OF ANY KIND, EITHER | |
20 | * EXPRESS OR IMPLIED, AND APPLE HEREBY DISCLAIMS ALL SUCH WARRANTIES, | |
21 | * INCLUDING WITHOUT LIMITATION, ANY WARRANTIES OF MERCHANTABILITY, | |
22 | * FITNESS FOR A PARTICULAR PURPOSE, QUIET ENJOYMENT OR NON-INFRINGEMENT. | |
23 | * Please see the License for the specific language governing rights and | |
24 | * limitations under the License. | |
25 | * | |
26 | * @APPLE_OSREFERENCE_LICENSE_HEADER_END@ | |
27 | */ | |
28 | /* | |
29 | * @OSF_COPYRIGHT@ | |
30 | * | |
31 | */ | |
32 | ||
33 | #include <debug.h> | |
34 | #include <mach/vm_param.h> | |
35 | #include <vm/vm_kern.h> | |
36 | #include <vm/vm_map.h> | |
37 | #include <vm/vm_page.h> | |
38 | #include <ppc/pmap.h> | |
39 | #include <ppc/io_map_entries.h> | |
40 | #include <ppc/Firmware.h> | |
41 | #include <ppc/mappings.h> | |
42 | #include <ppc/proc_reg.h> | |
43 | ||
44 | extern vm_offset_t virtual_avail; | |
45 | ||
46 | /* | |
47 | * Allocate and map memory for devices that may need to be mapped | |
48 | * outside the usual physical memory. If phys_addr is NULL then | |
49 | * steal the appropriate number of physical pages from the vm | |
50 | * system and map them. | |
51 | * | |
52 | * Note, this will onl | |
53 | */ | |
54 | vm_offset_t | |
55 | io_map(vm_offset_t phys_addr, vm_size_t size, unsigned int flags) | |
56 | { | |
57 | vm_offset_t start; | |
58 | vm_size_t i; | |
59 | unsigned int mflags; | |
60 | vm_page_t m; | |
61 | ||
62 | mflags = mmFlgBlock | mmFlgUseAttr | (flags & VM_MEM_GUARDED) | ((flags & VM_MEM_NOT_CACHEABLE) >> 1); /* Convert to our mapping_make flags */ | |
63 | ||
64 | #if DEBUG | |
65 | assert (kernel_map != VM_MAP_NULL); /* VM must be initialised */ | |
66 | #endif | |
67 | ||
68 | if (phys_addr != 0) { /* If they supplied a physical address, use it */ | |
69 | ||
70 | size = round_page(size + (phys_addr & PAGE_MASK)); /* Make sure we map all of it */ | |
71 | ||
72 | (void) kmem_alloc_pageable(kernel_map, &start, size); /* Get some virtual addresses to use */ | |
73 | ||
74 | (void)mapping_make(kernel_pmap, (addr64_t)start, (ppnum_t)(phys_addr >> 12), | |
75 | mflags, /* Map with requested cache mode */ | |
76 | (size >> 12), VM_PROT_READ|VM_PROT_WRITE); | |
77 | ||
78 | return (start + (phys_addr & PAGE_MASK)); /* Pass back the physical address */ | |
79 | ||
80 | } else { | |
81 | ||
82 | (void) kmem_alloc_pageable(kernel_map, &start, size); /* Get some virtual addresses */ | |
83 | ||
84 | mapping_prealloc(size); /* Make sure there are enough free mappings */ | |
85 | ||
86 | for (i = 0; i < size ; i += PAGE_SIZE) { | |
87 | m = VM_PAGE_NULL; | |
88 | while ((m = vm_page_grab()) == VM_PAGE_NULL) { /* Get a physical page */ | |
89 | VM_PAGE_WAIT(); /* Wait if we didn't have one */ | |
90 | } | |
91 | vm_page_gobble(m); | |
92 | ||
93 | (void)mapping_make(kernel_pmap, | |
94 | (addr64_t)(start + i), m->phys_page, | |
95 | mflags, /* Map with requested cache mode */ | |
96 | 1, VM_PROT_READ|VM_PROT_WRITE); | |
97 | ||
98 | } | |
99 | ||
100 | mapping_relpre(); /* Allow mapping release */ | |
101 | return start; | |
102 | } | |
103 | } | |
104 | ||
105 | ||
106 | /* | |
107 | * Allocate and map memory for devices before the VM system comes alive. | |
108 | */ | |
109 | ||
110 | vm_offset_t io_map_spec(vm_offset_t phys_addr, vm_size_t size, unsigned int flags) | |
111 | { | |
112 | vm_offset_t start; | |
113 | unsigned int mflags; | |
114 | ||
115 | if(kernel_map != VM_MAP_NULL) { /* If VM system is up, redirect to normal routine */ | |
116 | ||
117 | return io_map(phys_addr, size, flags); /* Map the address */ | |
118 | ||
119 | } | |
120 | ||
121 | mflags = mmFlgBlock | mmFlgUseAttr | (flags & VM_MEM_GUARDED) | ((flags & VM_MEM_NOT_CACHEABLE) >> 1); /* Convert to our mapping_make flags */ | |
122 | ||
123 | size = round_page(size + (phys_addr - (phys_addr & -PAGE_SIZE))); /* Extend the length to include it all */ | |
124 | start = pmap_boot_map(size); /* Get me some virtual address */ | |
125 | ||
126 | (void)mapping_make(kernel_pmap, (addr64_t)start, (ppnum_t)(phys_addr >> 12), | |
127 | mflags, /* Map with requested cache mode */ | |
128 | (size >> 12), VM_PROT_READ|VM_PROT_WRITE); | |
129 | ||
130 | return (start + (phys_addr & PAGE_MASK)); | |
131 | } |