]> git.saurik.com Git - apple/xnu.git/blob - osfmk/i386/start64.s
xnu-2050.22.13.tar.gz
[apple/xnu.git] / osfmk / i386 / start64.s
1 /*
2 * Copyright (c) 2006 Apple Computer, Inc. All rights reserved.
3 *
4 * @APPLE_OSREFERENCE_LICENSE_HEADER_START@
5 *
6 * This file contains Original Code and/or Modifications of Original Code
7 * as defined in and that are subject to the Apple Public Source License
8 * Version 2.0 (the 'License'). You may not use this file except in
9 * compliance with the License. The rights granted to you under the License
10 * may not be used to create, or enable the creation or redistribution of,
11 * unlawful or unlicensed copies of an Apple operating system, or to
12 * circumvent, violate, or enable the circumvention or violation of, any
13 * terms of an Apple operating system software license agreement.
14 *
15 * Please obtain a copy of the License at
16 * http://www.opensource.apple.com/apsl/ and read it before using this file.
17 *
18 * The Original Code and all software distributed under the License are
19 * distributed on an 'AS IS' basis, WITHOUT WARRANTY OF ANY KIND, EITHER
20 * EXPRESS OR IMPLIED, AND APPLE HEREBY DISCLAIMS ALL SUCH WARRANTIES,
21 * INCLUDING WITHOUT LIMITATION, ANY WARRANTIES OF MERCHANTABILITY,
22 * FITNESS FOR A PARTICULAR PURPOSE, QUIET ENJOYMENT OR NON-INFRINGEMENT.
23 * Please see the License for the specific language governing rights and
24 * limitations under the License.
25 *
26 * @APPLE_OSREFERENCE_LICENSE_HEADER_END@
27 */
28
29 #include <platforms.h>
30 #include <i386/asm.h>
31 #include <i386/asm64.h>
32 #include <i386/proc_reg.h>
33 #include <i386/postcode.h>
34 #include <i386/vmx/vmx_asm.h>
35 #include <assym.s>
36
37 .data
38 .align 3
39 .globl EXT(gdtptr64)
40 /* align below right */
41 .word 0
42 LEXT(gdtptr64)
43 .word Times(8,GDTSZ)-1
44 /* XXX really want .quad here */
45 .long EXT(master_gdt)
46 .long KERNEL_UBER_BASE_HI32 /* must be in uber-space */
47
48 .align 3
49 .globl EXT(idtptr64)
50 /* align below right */
51 .word 0
52 LEXT(idtptr64)
53 .word Times(16,IDTSZ)-1
54 /* XXX really want .quad here */
55 .long EXT(master_idt64)
56 .long KERNEL_UBER_BASE_HI32 /* must be in uber-space */
57
58 .text
59
60 Entry(ml_load_desc64)
61
62 ENTER_64BIT_MODE()
63
64 POSTCODE(ML_LOAD_DESC64_ENTRY)
65
66 lgdt EXT(gdtptr64) /* load GDT */
67
68 POSTCODE(ML_LOAD_DESC64_GDT)
69
70 lidt EXT(idtptr64) /* load IDT */
71
72 POSTCODE(ML_LOAD_DESC64_IDT)
73
74 movw $(KERNEL_LDT),%ax /* get LDT segment */
75 lldt %ax /* load LDT */
76
77 POSTCODE(ML_LOAD_DESC64_LDT)
78
79 movw $(KERNEL_TSS),%ax
80 ltr %ax /* set up KTSS */
81
82 POSTCODE(ML_LOAD_DESC64_EXIT)
83
84 ENTER_COMPAT_MODE()
85
86 ret
87
88
89 Entry(ml_64bit_lldt)
90 /* (int32_t selector) */
91
92 FRAME
93
94 ENTER_64BIT_MODE()
95
96 movl B_ARG0, %eax
97 lldt %ax
98
99 ENTER_COMPAT_MODE()
100
101 EMARF
102 ret
103
104 Entry(set_64bit_debug_regs)
105 /* x86_debug_state64_t *ds */
106
107 FRAME
108
109 ENTER_64BIT_MODE()
110
111 mov B_ARG0, %edx
112 mov DS64_DR0(%edx), %rax
113 mov %rax, %dr0
114 mov DS64_DR1(%edx), %rax
115 mov %rax, %dr1
116 mov DS64_DR2(%edx), %rax
117 mov %rax, %dr2
118 mov DS64_DR3(%edx), %rax
119 mov %rax, %dr3
120
121 ENTER_COMPAT_MODE()
122
123 EMARF
124 ret
125
126 Entry(flush_tlb64)
127
128 FRAME
129
130 ENTER_64BIT_MODE()
131
132 mov %cr3, %rax
133 mov %rax, %cr3
134
135 ENTER_COMPAT_MODE()
136
137 EMARF
138 ret
139
140 Entry(set64_cr3)
141
142 FRAME
143
144 movl B_ARG0, %eax
145 movl B_ARG1, %edx
146
147 ENTER_64BIT_MODE()
148
149 /* %rax = %edx:%eax */
150 shl $32, %rax
151 shrd $32, %rdx, %rax
152
153 mov %rax, %cr3
154
155 ENTER_COMPAT_MODE()
156
157 EMARF
158 ret
159
160 Entry(get64_cr3)
161
162 FRAME
163
164 ENTER_64BIT_MODE()
165
166 mov %cr3, %rax
167 mov %rax, %rdx
168 shr $32, %rdx // %edx:%eax = %cr3
169
170 ENTER_COMPAT_MODE()
171
172 EMARF
173 ret
174
175 Entry(cpuid64)
176 ENTER_64BIT_MODE()
177 cpuid
178 ENTER_COMPAT_MODE()
179 ret
180
181
182 /* FXSAVE and FXRSTOR operate in a mode dependent fashion, hence these variants.
183 * Must be called with interrupts disabled.
184 */
185
186 Entry(fxsave64)
187 movl S_ARG0,%eax
188 ENTER_64BIT_MODE()
189 fxsave (%eax)
190 ENTER_COMPAT_MODE()
191 ret
192
193 Entry(fxrstor64)
194 movl S_ARG0,%eax
195 ENTER_64BIT_MODE()
196 fxrstor (%rax)
197 ENTER_COMPAT_MODE()
198 ret
199
200 Entry(xsave64o)
201 ENTER_64BIT_MODE()
202 .short 0xAE0F
203 /* MOD 0x4, ECX, 0x1 */
204 .byte 0x21
205 ENTER_COMPAT_MODE()
206 ret
207
208 Entry(xrstor64o)
209 ENTER_64BIT_MODE()
210 .short 0xAE0F
211 /* MOD 0x5, ECX 0x1 */
212 .byte 0x29
213 ENTER_COMPAT_MODE()
214 ret
215
216 #if CONFIG_VMX
217
218 /*
219 * __vmxon -- Enter VMX Operation
220 * int __vmxon(addr64_t v);
221 */
222 Entry(__vmxon)
223 FRAME
224
225 ENTER_64BIT_MODE()
226 mov $(VMX_FAIL_INVALID), %ecx
227 mov $(VMX_FAIL_VALID), %edx
228 mov $(VMX_SUCCEED), %eax
229 vmxon 8(%rbp) /* physical addr passed on stack */
230 cmovcl %ecx, %eax /* CF = 1, ZF = 0 */
231 cmovzl %edx, %eax /* CF = 0, ZF = 1 */
232 ENTER_COMPAT_MODE()
233
234 EMARF
235 ret
236
237 /*
238 * __vmxoff -- Leave VMX Operation
239 * int __vmxoff(void);
240 */
241 Entry(__vmxoff)
242 FRAME
243
244 ENTER_64BIT_MODE()
245 mov $(VMX_FAIL_INVALID), %ecx
246 mov $(VMX_FAIL_VALID), %edx
247 mov $(VMX_SUCCEED), %eax
248 vmxoff
249 cmovcl %ecx, %eax /* CF = 1, ZF = 0 */
250 cmovzl %edx, %eax /* CF = 0, ZF = 1 */
251 ENTER_COMPAT_MODE()
252
253 EMARF
254 ret
255
256 #endif /* CONFIG_VMX */