]> git.saurik.com Git - apple/xnu.git/blame_incremental - osfmk/arm64/tlb.h
xnu-7195.101.1.tar.gz
[apple/xnu.git] / osfmk / arm64 / tlb.h
... / ...
CommitLineData
1/*
2 * Copyright (c) 2019 Apple Inc. All rights reserved.
3 *
4 * @APPLE_OSREFERENCE_LICENSE_HEADER_START@
5 *
6 * This file contains Original Code and/or Modifications of Original Code
7 * as defined in and that are subject to the Apple Public Source License
8 * Version 2.0 (the 'License'). You may not use this file except in
9 * compliance with the License. The rights granted to you under the License
10 * may not be used to create, or enable the creation or redistribution of,
11 * unlawful or unlicensed copies of an Apple operating system, or to
12 * circumvent, violate, or enable the circumvention or violation of, any
13 * terms of an Apple operating system software license agreement.
14 *
15 * Please obtain a copy of the License at
16 * http://www.opensource.apple.com/apsl/ and read it before using this file.
17 *
18 * The Original Code and all software distributed under the License are
19 * distributed on an 'AS IS' basis, WITHOUT WARRANTY OF ANY KIND, EITHER
20 * EXPRESS OR IMPLIED, AND APPLE HEREBY DISCLAIMS ALL SUCH WARRANTIES,
21 * INCLUDING WITHOUT LIMITATION, ANY WARRANTIES OF MERCHANTABILITY,
22 * FITNESS FOR A PARTICULAR PURPOSE, QUIET ENJOYMENT OR NON-INFRINGEMENT.
23 * Please see the License for the specific language governing rights and
24 * limitations under the License.
25 *
26 * @APPLE_OSREFERENCE_LICENSE_HEADER_END@
27 */
28
29#pragma once
30
31#include <arm64/proc_reg.h>
32#include <machine/atomic.h>
33
34#define tlbi_addr(x) ((((x) >> 12) & TLBI_ADDR_MASK) << TLBI_ADDR_SHIFT)
35#define tlbi_asid(x) (((uintptr_t)(x) & TLBI_ASID_MASK) << TLBI_ASID_SHIFT)
36
37#if __ARM_KERNEL_PROTECT__
38/*
39 * __ARM_KERNEL_PROTECT__ adds two complications to TLB management:
40 *
41 * 1. As each pmap has two ASIDs, every TLB operation that targets an ASID must
42 * target both ASIDs for the pmap that owns the target ASID.
43 *
44 * 2. Any TLB operation targeting the kernel_pmap ASID (ASID 0) must target all
45 * ASIDs (as kernel_pmap mappings may be referenced while using an ASID that
46 * belongs to another pmap). We expect these routines to be called with the
47 * EL0 ASID for the target; not the EL1 ASID.
48 */
49#endif /* __ARM_KERNEL_PROTECT__ */
50
51static inline void
52sync_tlb_flush(void)
53{
54 __builtin_arm_dsb(DSB_ISH);
55 __builtin_arm_isb(ISB_SY);
56}
57
58static inline void
59sync_tlb_flush_local(void)
60{
61 __builtin_arm_dsb(DSB_NSH);
62 __builtin_arm_isb(ISB_SY);
63}
64
65// flush_mmu_tlb: full TLB flush on all cores
66static inline void
67flush_mmu_tlb_async(void)
68{
69 asm volatile ("tlbi vmalle1is");
70}
71
72static inline void
73flush_mmu_tlb(void)
74{
75 flush_mmu_tlb_async();
76 sync_tlb_flush();
77}
78
79// flush_core_tlb: full TLB flush on local core only
80static inline void
81flush_core_tlb_async(void)
82{
83 asm volatile ("tlbi vmalle1");
84}
85
86static inline void
87flush_core_tlb(void)
88{
89 flush_core_tlb_async();
90 sync_tlb_flush();
91}
92
93// flush_mmu_tlb_allentries_async: flush entries that map VA range, all ASIDS, all cores
94// start and end are in units of 4K pages.
95static inline void
96flush_mmu_tlb_allentries_async(uint64_t start, uint64_t end, uint64_t pmap_page_size)
97{
98#if __ARM_16K_PG__
99 if (pmap_page_size == 16384) {
100 start = start & ~0x3ULL;
101
102 /*
103 * The code below is not necessarily correct. From an overview of
104 * the client code, the expected contract for TLB flushes is that
105 * we will expand from an "address, length" pair to "start address,
106 * end address" in the course of a TLB flush. This suggests that
107 * a flush for "X, X+4" is actually only asking for a flush of a
108 * single 16KB page. At the same time, we'd like to be prepared
109 * for bad inputs (X, X+3), so add 3 and then truncate the 4KB page
110 * number to a 16KB page boundary. This should deal correctly with
111 * unaligned inputs.
112 *
113 * If our expecations about client behavior are wrong however, this
114 * will lead to occasional TLB corruption on platforms with 16KB
115 * pages.
116 */
117 end = (end + 0x3ULL) & ~0x3ULL;
118 }
119#endif // __ARM_16K_PG__
120 for (; start < end; start += (pmap_page_size / 4096)) {
121 asm volatile ("tlbi vaae1is, %0" : : "r"(start));
122 }
123}
124
125static inline void
126flush_mmu_tlb_allentries(uint64_t start, uint64_t end, uint64_t pmap_page_size)
127{
128 flush_mmu_tlb_allentries_async(start, end, pmap_page_size);
129 sync_tlb_flush();
130}
131
132// flush_mmu_tlb_entry: flush TLB entries that map a VA and ASID, all cores
133// Will also flush global entries that match the VA
134static inline void
135flush_mmu_tlb_entry_async(uint64_t val)
136{
137#if __ARM_KERNEL_PROTECT__
138 uint64_t asid = val >> TLBI_ASID_SHIFT;
139 if (asid == 0) {
140 asm volatile ("tlbi vaae1is, %0" : : "r"(val));
141 return;
142 }
143 val = val & ~(1ULL << TLBI_ASID_SHIFT);
144 asm volatile ("tlbi vae1is, %0" : : "r"(val));
145 val = val | (1ULL << TLBI_ASID_SHIFT);
146#endif /* __ARM_KERNEL_PROTECT__ */
147 asm volatile ("tlbi vae1is, %0" : : "r"(val));
148}
149
150static inline void
151flush_mmu_tlb_entry(uint64_t val)
152{
153 flush_mmu_tlb_entry_async(val);
154 sync_tlb_flush();
155}
156
157// flush_mmu_tlb_entries: flush TLB entries that map a VA range and ASID, all cores
158// start and end must have the ASID in the high 16 bits, with the VA in units of 4K in the lowest bits
159// Will also flush global entries that match the VA range
160static inline void
161flush_mmu_tlb_entries_async(uint64_t start, uint64_t end, uint64_t pmap_page_size)
162{
163#if __ARM_16K_PG__
164 if (pmap_page_size == 16384) {
165 start = start & ~0x3ULL;
166
167 /*
168 * The code below is not necessarily correct. From an overview of
169 * the client code, the expected contract for TLB flushes is that
170 * we will expand from an "address, length" pair to "start address,
171 * end address" in the course of a TLB flush. This suggests that
172 * a flush for "X, X+4" is actually only asking for a flush of a
173 * single 16KB page. At the same time, we'd like to be prepared
174 * for bad inputs (X, X+3), so add 3 and then truncate the 4KB page
175 * number to a 16KB page boundary. This should deal correctly with
176 * unaligned inputs.
177 *
178 * If our expecations about client behavior are wrong however, this
179 * will lead to occasional TLB corruption on platforms with 16KB
180 * pages.
181 */
182 end = (end + 0x3ULL) & ~0x3ULL;
183 }
184#endif // __ARM_16K_PG__
185#if __ARM_KERNEL_PROTECT__
186 uint64_t asid = start >> TLBI_ASID_SHIFT;
187 /*
188 * If we are flushing ASID 0, this is a kernel operation. With this
189 * ASID scheme, this means we should flush all ASIDs.
190 */
191 if (asid == 0) {
192 for (; start < end; start += (pmap_page_size / 4096)) {
193 asm volatile ("tlbi vaae1is, %0" : : "r"(start));
194 }
195 return;
196 }
197 start = start | (1ULL << TLBI_ASID_SHIFT);
198 end = end | (1ULL << TLBI_ASID_SHIFT);
199 for (; start < end; start += (pmap_page_size / 4096)) {
200 start = start & ~(1ULL << TLBI_ASID_SHIFT);
201 asm volatile ("tlbi vae1is, %0" : : "r"(start));
202 start = start | (1ULL << TLBI_ASID_SHIFT);
203 asm volatile ("tlbi vae1is, %0" : : "r"(start));
204 }
205#else
206 for (; start < end; start += (pmap_page_size / 4096)) {
207 asm volatile ("tlbi vae1is, %0" : : "r"(start));
208 }
209#endif /* __ARM_KERNEL_PROTECT__ */
210}
211
212static inline void
213flush_mmu_tlb_entries(uint64_t start, uint64_t end, uint64_t pmap_page_size)
214{
215 flush_mmu_tlb_entries_async(start, end, pmap_page_size);
216 sync_tlb_flush();
217}
218
219// flush_mmu_tlb_asid: flush all entries that match an ASID, on all cores
220// ASID must be in high 16 bits of argument
221// Will not flush global entries
222static inline void
223flush_mmu_tlb_asid_async(uint64_t val)
224{
225#if __ARM_KERNEL_PROTECT__
226 /*
227 * If we are flushing ASID 0, this is a kernel operation. With this
228 * ASID scheme, this means we should flush all ASIDs.
229 */
230 uint64_t asid = val >> TLBI_ASID_SHIFT;
231 if (asid == 0) {
232 asm volatile ("tlbi vmalle1is");
233 return;
234 }
235 val = val & ~(1ULL << TLBI_ASID_SHIFT);
236 asm volatile ("tlbi aside1is, %0" : : "r"(val));
237 val = val | (1ULL << TLBI_ASID_SHIFT);
238#endif /* __ARM_KERNEL_PROTECT__ */
239 asm volatile ("tlbi aside1is, %0" : : "r"(val));
240}
241
242static inline void
243flush_mmu_tlb_asid(uint64_t val)
244{
245 flush_mmu_tlb_asid_async(val);
246 sync_tlb_flush();
247}
248
249// flush_core_tlb_asid: flush all entries that match an ASID, local core only
250// ASID must be in high 16 bits of argument
251// Will not flush global entries
252static inline void
253flush_core_tlb_asid_async(uint64_t val)
254{
255#if __ARM_KERNEL_PROTECT__
256 /*
257 * If we are flushing ASID 0, this is a kernel operation. With this
258 * ASID scheme, this means we should flush all ASIDs.
259 */
260 uint64_t asid = val >> TLBI_ASID_SHIFT;
261 if (asid == 0) {
262 asm volatile ("tlbi vmalle1");
263 return;
264 }
265 val = val & ~(1ULL << TLBI_ASID_SHIFT);
266 asm volatile ("tlbi aside1, %0" : : "r"(val));
267 val = val | (1ULL << TLBI_ASID_SHIFT);
268#endif /* __ARM_KERNEL_PROTECT__ */
269 asm volatile ("tlbi aside1, %0" : : "r"(val));
270}
271
272static inline void
273flush_core_tlb_asid(uint64_t val)
274{
275 flush_core_tlb_asid_async(val);
276 sync_tlb_flush();
277}
278
279#if __ARM_RANGE_TLBI__
280#if __ARM_KERNEL_PROTECT__
281 #error __ARM_RANGE_TLBI__ + __ARM_KERNEL_PROTECT__ is not currently supported
282#endif
283
284#define ARM64_TLB_RANGE_PAGES (1ULL << 21)
285#define rtlbi_addr(x, shift) (((x) >> (shift)) & RTLBI_ADDR_MASK)
286#define rtlbi_scale(x) ((uint64_t)(x) << RTLBI_SCALE_SHIFT)
287#define rtlbi_num(x) ((uint64_t)(x) << RTLBI_NUM_SHIFT)
288
289/**
290 * Given the number of pages to invalidate, generate the correct parameter to
291 * pass to any of the TLBI by range methods.
292 */
293static inline uint64_t
294generate_rtlbi_param(ppnum_t npages, uint32_t asid, vm_offset_t va, uint64_t pmap_page_shift)
295{
296 /**
297 * Per the armv8.4 RTLBI extension spec, the range encoded in the rtlbi register operand is defined by:
298 * BaseADDR <= VA < BaseADDR+((NUM+1)*2^(5*SCALE+1) * Translation_Granule_Size)
299 */
300 unsigned order = (unsigned)(sizeof(npages) * 8) - (unsigned)__builtin_clz(npages - 1) - 1;
301 unsigned scale = ((order ? order : 1) - 1) / 5;
302 unsigned granule = 1 << ((5 * scale) + 1);
303 unsigned num = (((npages + granule - 1) & ~(granule - 1)) / granule) - 1;
304 return tlbi_asid(asid) | RTLBI_TG(pmap_page_shift) | rtlbi_scale(scale) | rtlbi_num(num) | rtlbi_addr(va, pmap_page_shift);
305}
306
307// flush_mmu_tlb_range: flush TLB entries that map a VA range using a single instruction
308// The argument should be encoded according to generate_rtlbi_param().
309// Follows the same ASID matching behavior as flush_mmu_tlb_entries()
310static inline void
311flush_mmu_tlb_range_async(uint64_t val)
312{
313 asm volatile ("tlbi rvae1is, %0" : : "r"(val));
314}
315
316static inline void
317flush_mmu_tlb_range(uint64_t val)
318{
319 flush_mmu_tlb_range_async(val);
320 sync_tlb_flush();
321}
322
323// flush_mmu_tlb_allrange: flush TLB entries that map a VA range using a single instruction
324// The argument should be encoded according to generate_rtlbi_param().
325// Follows the same ASID matching behavior as flush_mmu_tlb_allentries()
326static inline void
327flush_mmu_tlb_allrange_async(uint64_t val)
328{
329 asm volatile ("tlbi rvaae1is, %0" : : "r"(val));
330}
331
332static inline void
333flush_mmu_tlb_allrange(uint64_t val)
334{
335 flush_mmu_tlb_allrange_async(val);
336 sync_tlb_flush();
337}
338
339// flush_core_tlb_allrange: flush TLB entries that map a VA range using a single instruction, local core only
340// The argument should be encoded according to generate_rtlbi_param().
341// Follows the same ASID matching behavior as flush_mmu_tlb_allentries()
342static inline void
343flush_core_tlb_allrange_async(uint64_t val)
344{
345 asm volatile ("tlbi rvaae1, %0" : : "r"(val));
346}
347
348static inline void
349flush_core_tlb_allrange(uint64_t val)
350{
351 flush_core_tlb_allrange_async(val);
352 sync_tlb_flush();
353}
354
355#endif // __ARM_RANGE_TLBI__
356
357