]> git.saurik.com Git - apple/xnu.git/blame_incremental - osfmk/ppc/ppc_vm_init.c
xnu-792.13.8.tar.gz
[apple/xnu.git] / osfmk / ppc / ppc_vm_init.c
... / ...
CommitLineData
1/*
2 * Copyright (c) 2000-2005 Apple Computer, Inc. All rights reserved.
3 *
4 * @APPLE_LICENSE_OSREFERENCE_HEADER_START@
5 *
6 * This file contains Original Code and/or Modifications of Original Code
7 * as defined in and that are subject to the Apple Public Source License
8 * Version 2.0 (the 'License'). You may not use this file except in
9 * compliance with the License. The rights granted to you under the
10 * License may not be used to create, or enable the creation or
11 * redistribution of, unlawful or unlicensed copies of an Apple operating
12 * system, or to circumvent, violate, or enable the circumvention or
13 * violation of, any terms of an Apple operating system software license
14 * agreement.
15 *
16 * Please obtain a copy of the License at
17 * http://www.opensource.apple.com/apsl/ and read it before using this
18 * file.
19 *
20 * The Original Code and all software distributed under the License are
21 * distributed on an 'AS IS' basis, WITHOUT WARRANTY OF ANY KIND, EITHER
22 * EXPRESS OR IMPLIED, AND APPLE HEREBY DISCLAIMS ALL SUCH WARRANTIES,
23 * INCLUDING WITHOUT LIMITATION, ANY WARRANTIES OF MERCHANTABILITY,
24 * FITNESS FOR A PARTICULAR PURPOSE, QUIET ENJOYMENT OR NON-INFRINGEMENT.
25 * Please see the License for the specific language governing rights and
26 * limitations under the License.
27 *
28 * @APPLE_LICENSE_OSREFERENCE_HEADER_END@
29 */
30/*
31 * @OSF_COPYRIGHT@
32 */
33/*
34 * @APPLE_FREE_COPYRIGHT@
35 */
36
37#include <mach_debug.h>
38#include <mach_kdb.h>
39#include <mach_kdp.h>
40#include <debug.h>
41
42#include <mach/vm_types.h>
43#include <mach/vm_param.h>
44#include <mach/thread_status.h>
45#include <kern/misc_protos.h>
46#include <kern/assert.h>
47#include <kern/cpu_number.h>
48#include <kern/thread.h>
49#include <console/serial_protos.h>
50
51#include <ppc/proc_reg.h>
52#include <ppc/Firmware.h>
53#include <ppc/boot.h>
54#include <ppc/misc_protos.h>
55#include <ppc/pmap.h>
56#include <ppc/mem.h>
57#include <ppc/mappings.h>
58#include <ppc/exception.h>
59#include <ppc/lowglobals.h>
60
61#include <mach-o/mach_header.h>
62
63extern const char version[];
64extern const char version_variant[];
65
66addr64_t hash_table_base; /* Hash table base */
67unsigned int hash_table_size; /* Hash table size */
68int hash_table_shift; /* "ht_shift" boot arg, used to scale hash_table_size */
69vm_offset_t taproot_addr; /* (BRINGUP) */
70unsigned int taproot_size; /* (BRINGUP) */
71extern int disableConsoleOutput;
72
73struct shadowBAT shadow_BAT;
74
75
76
77/*
78 * NOTE: mem_size is bogus on large memory machines. We will pin it to 0x80000000 if there is more than 2 GB
79 * This is left only for compatibility and max_mem should be used.
80 */
81vm_offset_t mem_size; /* Size of actual physical memory present
82 minus any performance buffer and possibly limited
83 by mem_limit in bytes */
84uint64_t mem_actual; /* The "One True" physical memory size
85 actually, it's the highest physical address + 1 */
86uint64_t max_mem; /* Size of physical memory (bytes), adjusted by maxmem */
87uint64_t sane_size; /* Memory size to use for defaults calculations */
88
89
90mem_region_t pmap_mem_regions[PMAP_MEM_REGION_MAX + 1];
91int pmap_mem_regions_count = 0; /* Assume no non-contiguous memory regions */
92
93unsigned int avail_remaining = 0;
94vm_offset_t first_avail;
95vm_offset_t static_memory_end;
96addr64_t vm_last_addr = VM_MAX_KERNEL_ADDRESS; /* Highest kernel virtual address known to the VM system */
97
98extern struct mach_header _mh_execute_header;
99vm_offset_t sectTEXTB;
100int sectSizeTEXT;
101vm_offset_t sectDATAB;
102int sectSizeDATA;
103vm_offset_t sectLINKB;
104int sectSizeLINK;
105vm_offset_t sectKLDB;
106int sectSizeKLD;
107vm_offset_t sectPRELINKB;
108int sectSizePRELINK;
109vm_offset_t sectHIBB;
110int sectSizeHIB;
111
112vm_offset_t end, etext, edata;
113
114extern unsigned long exception_entry;
115extern unsigned long exception_end;
116
117
118void ppc_vm_init(uint64_t mem_limit, boot_args *args)
119{
120 unsigned int i, kmapsize, pvr;
121 vm_offset_t addr;
122 unsigned int *xtaproot, bank_shift;
123 uint64_t cbsize, xhid0;
124
125
126/*
127 * Invalidate all shadow BATs
128 */
129
130 /* Initialize shadow IBATs */
131 shadow_BAT.IBATs[0].upper=BAT_INVALID;
132 shadow_BAT.IBATs[0].lower=BAT_INVALID;
133 shadow_BAT.IBATs[1].upper=BAT_INVALID;
134 shadow_BAT.IBATs[1].lower=BAT_INVALID;
135 shadow_BAT.IBATs[2].upper=BAT_INVALID;
136 shadow_BAT.IBATs[2].lower=BAT_INVALID;
137 shadow_BAT.IBATs[3].upper=BAT_INVALID;
138 shadow_BAT.IBATs[3].lower=BAT_INVALID;
139
140 /* Initialize shadow DBATs */
141 shadow_BAT.DBATs[0].upper=BAT_INVALID;
142 shadow_BAT.DBATs[0].lower=BAT_INVALID;
143 shadow_BAT.DBATs[1].upper=BAT_INVALID;
144 shadow_BAT.DBATs[1].lower=BAT_INVALID;
145 shadow_BAT.DBATs[2].upper=BAT_INVALID;
146 shadow_BAT.DBATs[2].lower=BAT_INVALID;
147 shadow_BAT.DBATs[3].upper=BAT_INVALID;
148 shadow_BAT.DBATs[3].lower=BAT_INVALID;
149
150
151 /*
152 * Go through the list of memory regions passed in via the boot_args
153 * and copy valid entries into the pmap_mem_regions table, adding
154 * further calculated entries.
155 *
156 * boot_args version 1 has address instead of page numbers
157 * in the PhysicalDRAM banks, set bank_shift accordingly.
158 */
159
160 bank_shift = 0;
161 if (args->Version == kBootArgsVersion1) bank_shift = 12;
162
163 pmap_mem_regions_count = 0;
164 max_mem = 0; /* Will use to total memory found so far */
165 mem_actual = 0; /* Actual size of memory */
166
167 if (mem_limit == 0) mem_limit = 0xFFFFFFFFFFFFFFFFULL; /* If there is no set limit, use all */
168
169 for (i = 0; i < kMaxDRAMBanks; i++) { /* Look at all of the banks */
170
171 cbsize = (uint64_t)args->PhysicalDRAM[i].size << (12 - bank_shift); /* Remember current size */
172
173 if (!cbsize) continue; /* Skip if the bank is empty */
174
175 mem_actual = mem_actual + cbsize; /* Get true memory size */
176
177 if(mem_limit == 0) continue; /* If we hit restriction, just keep counting */
178
179 if (cbsize > mem_limit) cbsize = mem_limit; /* Trim to max allowed */
180 max_mem += cbsize; /* Total up what we have so far */
181 mem_limit = mem_limit - cbsize; /* Calculate amount left to do */
182
183 pmap_mem_regions[pmap_mem_regions_count].mrStart = args->PhysicalDRAM[i].base >> bank_shift; /* Set the start of the bank */
184 pmap_mem_regions[pmap_mem_regions_count].mrAStart = pmap_mem_regions[pmap_mem_regions_count].mrStart; /* Set the start of allocatable area */
185 pmap_mem_regions[pmap_mem_regions_count].mrEnd = ((uint64_t)args->PhysicalDRAM[i].base >> bank_shift) + (cbsize >> 12) - 1; /* Set the end address of bank */
186 pmap_mem_regions[pmap_mem_regions_count].mrAEnd = pmap_mem_regions[pmap_mem_regions_count].mrEnd; /* Set the end address of allocatable area */
187
188 /* Regions must be provided in ascending order */
189 assert ((pmap_mem_regions_count == 0) ||
190 pmap_mem_regions[pmap_mem_regions_count].mrStart >
191 pmap_mem_regions[pmap_mem_regions_count-1].mrStart);
192
193 pmap_mem_regions_count++; /* Count this region */
194 }
195
196 mem_size = (unsigned int)max_mem; /* Get size of memory */
197 if(max_mem > 0x0000000080000000ULL) mem_size = 0x80000000; /* Pin at 2 GB */
198
199 sane_size = max_mem; /* Calculate a sane value to use for init */
200 if(sane_size > (addr64_t)(VM_MAX_KERNEL_ADDRESS + 1))
201 sane_size = (addr64_t)(VM_MAX_KERNEL_ADDRESS + 1); /* If flush with ram, use addressible portion */
202
203
204/*
205 * Initialize the pmap system, using space above `first_avail'
206 * for the necessary data structures.
207 * NOTE : assume that we'll have enough space mapped in already
208 */
209
210 first_avail = static_memory_end;
211
212/*
213 * Now retrieve addresses for end, edata, and etext
214 * from MACH-O headers for the currently running 32 bit kernel.
215 */
216 sectTEXTB = (vm_offset_t)getsegdatafromheader(
217 &_mh_execute_header, "__TEXT", &sectSizeTEXT);
218 sectDATAB = (vm_offset_t)getsegdatafromheader(
219 &_mh_execute_header, "__DATA", &sectSizeDATA);
220 sectLINKB = (vm_offset_t)getsegdatafromheader(
221 &_mh_execute_header, "__LINKEDIT", &sectSizeLINK);
222 sectKLDB = (vm_offset_t)getsegdatafromheader(
223 &_mh_execute_header, "__KLD", &sectSizeKLD);
224 sectHIBB = (vm_offset_t)getsegdatafromheader(
225 &_mh_execute_header, "__HIB", &sectSizeHIB);
226 sectPRELINKB = (vm_offset_t)getsegdatafromheader(
227 &_mh_execute_header, "__PRELINK", &sectSizePRELINK);
228
229 etext = (vm_offset_t) sectTEXTB + sectSizeTEXT;
230 edata = (vm_offset_t) sectDATAB + sectSizeDATA;
231 end = round_page(getlastaddr()); /* Force end to next page */
232
233 kmapsize = (round_page(exception_end) - trunc_page(exception_entry)) + /* Get size we will map later */
234 (round_page(sectTEXTB+sectSizeTEXT) - trunc_page(sectTEXTB)) +
235 (round_page(sectDATAB+sectSizeDATA) - trunc_page(sectDATAB)) +
236 (round_page(sectLINKB+sectSizeLINK) - trunc_page(sectLINKB)) +
237 (round_page(sectKLDB+sectSizeKLD) - trunc_page(sectKLDB)) +
238 (round_page_32(sectKLDB+sectSizeHIB) - trunc_page_32(sectHIBB)) +
239 (round_page(sectPRELINKB+sectSizePRELINK) - trunc_page(sectPRELINKB)) +
240 (round_page(static_memory_end) - trunc_page(end));
241
242 pmap_bootstrap(max_mem, &first_avail, kmapsize);
243
244 pmap_map(trunc_page(exception_entry), trunc_page(exception_entry),
245 round_page(exception_end), VM_PROT_READ|VM_PROT_EXECUTE, VM_WIMG_USE_DEFAULT);
246
247 pmap_map(trunc_page(sectTEXTB), trunc_page(sectTEXTB),
248 round_page(sectTEXTB+sectSizeTEXT), VM_PROT_READ|VM_PROT_EXECUTE, VM_WIMG_USE_DEFAULT);
249
250 pmap_map(trunc_page(sectDATAB), trunc_page(sectDATAB),
251 round_page(sectDATAB+sectSizeDATA), VM_PROT_READ|VM_PROT_WRITE, VM_WIMG_USE_DEFAULT);
252
253/* The KLD and LINKEDIT segments are unloaded in toto after boot completes,
254* but via ml_static_mfree(), through IODTFreeLoaderInfo(). Hence, we have
255* to map both segments page-by-page.
256*/
257
258 for (addr = trunc_page(sectPRELINKB);
259 addr < round_page(sectPRELINKB+sectSizePRELINK);
260 addr += PAGE_SIZE) {
261
262 pmap_enter(kernel_pmap, (vm_map_offset_t)addr, (ppnum_t)(addr>>12),
263 VM_PROT_READ|VM_PROT_WRITE|VM_PROT_EXECUTE,
264 VM_WIMG_USE_DEFAULT, TRUE);
265
266 }
267
268 for (addr = trunc_page(sectKLDB);
269 addr < round_page(sectKLDB+sectSizeKLD);
270 addr += PAGE_SIZE) {
271
272 pmap_enter(kernel_pmap, (vm_map_offset_t)addr, (ppnum_t)(addr>>12),
273 VM_PROT_READ|VM_PROT_WRITE|VM_PROT_EXECUTE,
274 VM_WIMG_USE_DEFAULT, TRUE);
275
276 }
277
278 for (addr = trunc_page(sectLINKB);
279 addr < round_page(sectLINKB+sectSizeLINK);
280 addr += PAGE_SIZE) {
281
282 pmap_enter(kernel_pmap, (vm_map_offset_t)addr,
283 (ppnum_t)(addr>>12),
284 VM_PROT_READ|VM_PROT_WRITE|VM_PROT_EXECUTE,
285 VM_WIMG_USE_DEFAULT, TRUE);
286
287 }
288
289 for (addr = trunc_page_32(sectHIBB);
290 addr < round_page_32(sectHIBB+sectSizeHIB);
291 addr += PAGE_SIZE) {
292
293 pmap_enter(kernel_pmap, (vm_map_offset_t)addr, (ppnum_t)(addr>>12),
294 VM_PROT_READ|VM_PROT_WRITE|VM_PROT_EXECUTE,
295 VM_WIMG_USE_DEFAULT, TRUE);
296
297 }
298
299 pmap_enter(kernel_pmap, (vm_map_offset_t)&sharedPage,
300 (ppnum_t)&sharedPage >> 12, /* Make sure the sharedPage is mapped */
301 VM_PROT_READ|VM_PROT_WRITE,
302 VM_WIMG_USE_DEFAULT, TRUE);
303
304 pmap_enter(kernel_pmap, (vm_map_offset_t)&lowGlo.lgVerCode,
305 (ppnum_t)&lowGlo.lgVerCode >> 12, /* Make sure the low memory globals are mapped */
306 VM_PROT_READ|VM_PROT_WRITE,
307 VM_WIMG_USE_DEFAULT, TRUE);
308
309/*
310 * We need to map the remainder page-by-page because some of this will
311 * be released later, but not all. Ergo, no block mapping here
312 */
313
314 for(addr = trunc_page(end); addr < round_page(static_memory_end); addr += PAGE_SIZE) {
315
316 pmap_enter(kernel_pmap, (vm_map_address_t)addr, (ppnum_t)addr>>12,
317 VM_PROT_READ|VM_PROT_WRITE|VM_PROT_EXECUTE,
318 VM_WIMG_USE_DEFAULT, TRUE);
319
320 }
321
322/*
323 * Here we map a window into the kernel address space that will be used to
324 * access a slice of a user address space. Clients for this service include
325 * copyin/out and copypv.
326 */
327
328 lowGlo.lgUMWvaddr = USER_MEM_WINDOW_VADDR;
329 /* Initialize user memory window base address */
330 MapUserMemoryWindowInit(); /* Go initialize user memory window */
331
332/*
333 * At this point, there is enough mapped memory and all hw mapping structures are
334 * allocated and initialized. Here is where we turn on translation for the
335 * VERY first time....
336 *
337 * NOTE: Here is where our very first interruption will happen.
338 *
339 */
340
341 hw_start_trans(); /* Start translating */
342 PE_init_platform(TRUE, args); /* Initialize this right off the bat */
343
344
345#if 0
346 GratefulDebInit((bootBumbleC *)&(args->Video)); /* Initialize the GratefulDeb debugger */
347#endif
348
349
350 printf_init(); /* Init this in case we need debugger */
351 panic_init(); /* Init this in case we need debugger */
352 PE_init_kprintf(TRUE); /* Note on PPC we only call this after VM is set up */
353
354 kprintf("kprintf initialized\n");
355
356 serialmode = 0; /* Assume normal keyboard and console */
357 if(PE_parse_boot_arg("serial", &serialmode)) { /* Do we want a serial keyboard and/or console? */
358 kprintf("Serial mode specified: %08X\n", serialmode);
359 }
360 if(serialmode & 1) { /* Start serial if requested */
361 (void)switch_to_serial_console(); /* Switch into serial mode */
362 disableConsoleOutput = FALSE; /* Allow printfs to happen */
363 }
364
365 kprintf("max_mem: %ld M\n", (unsigned long)(max_mem >> 20));
366 kprintf("version_variant = %s\n", version_variant);
367 kprintf("version = %s\n\n", version);
368 __asm__ ("mfpvr %0" : "=r" (pvr));
369 kprintf("proc version = %08x\n", pvr);
370 if(getPerProc()->pf.Available & pf64Bit) { /* 64-bit processor? */
371 xhid0 = hid0get64(); /* Get the hid0 */
372 if(xhid0 & (1ULL << (63 - 19))) kprintf("Time base is externally clocked\n");
373 else kprintf("Time base is internally clocked\n");
374 }
375
376
377 taproot_size = PE_init_taproot(&taproot_addr); /* (BRINGUP) See if there is a taproot */
378 if(taproot_size) { /* (BRINGUP) */
379 kprintf("TapRoot card configured to use vaddr = %08X, size = %08X\n", taproot_addr, taproot_size);
380 bcopy_nc((void *)version, (void *)(taproot_addr + 16), strlen(version)); /* (BRINGUP) Pass it our kernel version */
381 __asm__ volatile("eieio"); /* (BRINGUP) */
382 xtaproot = (unsigned int *)taproot_addr; /* (BRINGUP) */
383 xtaproot[0] = 1; /* (BRINGUP) */
384 __asm__ volatile("eieio"); /* (BRINGUP) */
385 }
386
387 PE_create_console(); /* create the console for verbose or pretty mode */
388
389 /* setup console output */
390 PE_init_printf(FALSE);
391
392#if DEBUG
393 printf("\n\n\nThis program was compiled using gcc %d.%d for powerpc\n",
394 __GNUC__,__GNUC_MINOR__);
395
396
397 /* Processor version information */
398 {
399 unsigned int pvr;
400 __asm__ ("mfpvr %0" : "=r" (pvr));
401 printf("processor version register : %08X\n", pvr);
402 }
403
404 kprintf("Args at %08X\n", args);
405 for (i = 0; i < pmap_mem_regions_count; i++) {
406 printf("DRAM at %08X size %08X\n",
407 args->PhysicalDRAM[i].base,
408 args->PhysicalDRAM[i].size);
409 }
410#endif /* DEBUG */
411
412#if DEBUG
413 kprintf("Mapped memory:\n");
414 kprintf(" exception vector: %08X, %08X - %08X\n", trunc_page(exception_entry),
415 trunc_page(exception_entry), round_page(exception_end));
416 kprintf(" sectTEXTB: %08X, %08X - %08X\n", trunc_page(sectTEXTB),
417 trunc_page(sectTEXTB), round_page(sectTEXTB+sectSizeTEXT));
418 kprintf(" sectDATAB: %08X, %08X - %08X\n", trunc_page(sectDATAB),
419 trunc_page(sectDATAB), round_page(sectDATAB+sectSizeDATA));
420 kprintf(" sectLINKB: %08X, %08X - %08X\n", trunc_page(sectLINKB),
421 trunc_page(sectLINKB), round_page(sectLINKB+sectSizeLINK));
422 kprintf(" sectKLDB: %08X, %08X - %08X\n", trunc_page(sectKLDB),
423 trunc_page(sectKLDB), round_page(sectKLDB+sectSizeKLD));
424 kprintf(" end: %08X, %08X - %08X\n", trunc_page(end),
425 trunc_page(end), static_memory_end);
426
427#endif
428
429 return;
430}
431