]> git.saurik.com Git - apple/xnu.git/blob - osfmk/i386/Diagnostics.c
xnu-4570.51.1.tar.gz
[apple/xnu.git] / osfmk / i386 / Diagnostics.c
1 /*
2 * Copyright (c) 2005-2008 Apple Inc. All rights reserved.
3 *
4 * @APPLE_OSREFERENCE_LICENSE_HEADER_START@
5 *
6 * This file contains Original Code and/or Modifications of Original Code
7 * as defined in and that are subject to the Apple Public Source License
8 * Version 2.0 (the 'License'). You may not use this file except in
9 * compliance with the License. The rights granted to you under the License
10 * may not be used to create, or enable the creation or redistribution of,
11 * unlawful or unlicensed copies of an Apple operating system, or to
12 * circumvent, violate, or enable the circumvention or violation of, any
13 * terms of an Apple operating system software license agreement.
14 *
15 * Please obtain a copy of the License at
16 * http://www.opensource.apple.com/apsl/ and read it before using this file.
17 *
18 * The Original Code and all software distributed under the License are
19 * distributed on an 'AS IS' basis, WITHOUT WARRANTY OF ANY KIND, EITHER
20 * EXPRESS OR IMPLIED, AND APPLE HEREBY DISCLAIMS ALL SUCH WARRANTIES,
21 * INCLUDING WITHOUT LIMITATION, ANY WARRANTIES OF MERCHANTABILITY,
22 * FITNESS FOR A PARTICULAR PURPOSE, QUIET ENJOYMENT OR NON-INFRINGEMENT.
23 * Please see the License for the specific language governing rights and
24 * limitations under the License.
25 *
26 * @APPLE_OSREFERENCE_LICENSE_HEADER_END@
27 */
28 /*
29 * @OSF_FREE_COPYRIGHT@
30 */
31 /*
32 * @APPLE_FREE_COPYRIGHT@
33 */
34
35 /*
36 * Author: Bill Angell, Apple
37 * Date: 10/auht-five
38 *
39 * Random diagnostics, augmented Derek Kumar 2011
40 *
41 *
42 */
43
44
45 #include <kern/machine.h>
46 #include <kern/processor.h>
47 #include <mach/machine.h>
48 #include <mach/processor_info.h>
49 #include <mach/mach_types.h>
50 #include <mach/boolean.h>
51 #include <kern/thread.h>
52 #include <kern/task.h>
53 #include <kern/ipc_kobject.h>
54 #include <mach/vm_param.h>
55 #include <ipc/port.h>
56 #include <ipc/ipc_entry.h>
57 #include <ipc/ipc_space.h>
58 #include <ipc/ipc_object.h>
59 #include <ipc/ipc_port.h>
60 #include <vm/vm_kern.h>
61 #include <vm/vm_map.h>
62 #include <vm/vm_page.h>
63 #include <vm/pmap.h>
64 #include <pexpert/pexpert.h>
65 #include <console/video_console.h>
66 #include <i386/cpu_data.h>
67 #include <i386/Diagnostics.h>
68 #include <i386/mp.h>
69 #include <i386/pmCPU.h>
70 #include <i386/tsc.h>
71 #include <mach/i386/syscall_sw.h>
72 #include <kern/kalloc.h>
73 #include <sys/kdebug.h>
74 #include <i386/machine_cpu.h>
75 #include <i386/misc_protos.h>
76 #include <i386/cpuid.h>
77
78 #if MONOTONIC
79 #include <kern/monotonic.h>
80 #endif /* MONOTONIC */
81
82 #define PERMIT_PERMCHECK (0)
83
84 diagWork dgWork;
85 uint64_t lastRuptClear = 0ULL;
86 boolean_t diag_pmc_enabled = FALSE;
87 void cpu_powerstats(void *);
88
89 typedef struct {
90 uint64_t caperf;
91 uint64_t cmperf;
92 uint64_t ccres[6];
93 uint64_t crtimes[CPU_RTIME_BINS];
94 uint64_t citimes[CPU_ITIME_BINS];
95 uint64_t crtime_total;
96 uint64_t citime_total;
97 uint64_t cpu_idle_exits;
98 uint64_t cpu_insns;
99 uint64_t cpu_ucc;
100 uint64_t cpu_urc;
101 #if DIAG_ALL_PMCS
102 uint64_t gpmcs[4];
103 #endif /* DIAG_ALL_PMCS */
104 } core_energy_stat_t;
105
106 typedef struct {
107 uint64_t pkes_version;
108 uint64_t pkg_cres[2][7];
109 uint64_t pkg_power_unit;
110 uint64_t pkg_energy;
111 uint64_t pp0_energy;
112 uint64_t pp1_energy;
113 uint64_t ddr_energy;
114 uint64_t llc_flushed_cycles;
115 uint64_t ring_ratio_instantaneous;
116 uint64_t IA_frequency_clipping_cause;
117 uint64_t GT_frequency_clipping_cause;
118 uint64_t pkg_idle_exits;
119 uint64_t pkg_rtimes[CPU_RTIME_BINS];
120 uint64_t pkg_itimes[CPU_ITIME_BINS];
121 uint64_t mbus_delay_time;
122 uint64_t mint_delay_time;
123 uint32_t ncpus;
124 core_energy_stat_t cest[];
125 } pkg_energy_statistics_t;
126
127
128 int
129 diagCall64(x86_saved_state_t * state)
130 {
131 uint64_t curpos, i, j;
132 uint64_t selector, data;
133 uint64_t currNap, durNap;
134 x86_saved_state64_t *regs;
135 boolean_t diagflag;
136 uint32_t rval = 0;
137
138 assert(is_saved_state64(state));
139 regs = saved_state64(state);
140
141 diagflag = ((dgWork.dgFlags & enaDiagSCs) != 0);
142 selector = regs->rdi;
143
144 switch (selector) { /* Select the routine */
145 case dgRuptStat: /* Suck Interruption statistics */
146 (void) ml_set_interrupts_enabled(TRUE);
147 data = regs->rsi; /* Get the number of processors */
148
149 if (data == 0) { /* If no location is specified for data, clear all
150 * counts
151 */
152 for (i = 0; i < real_ncpus; i++) { /* Cycle through
153 * processors */
154 for (j = 0; j < 256; j++)
155 cpu_data_ptr[i]->cpu_hwIntCnt[j] = 0;
156 }
157
158 lastRuptClear = mach_absolute_time(); /* Get the time of clear */
159 rval = 1; /* Normal return */
160 (void) ml_set_interrupts_enabled(FALSE);
161 break;
162 }
163
164 (void) copyout((char *) &real_ncpus, data, sizeof(real_ncpus)); /* Copy out number of
165 * processors */
166 currNap = mach_absolute_time(); /* Get the time now */
167 durNap = currNap - lastRuptClear; /* Get the last interval
168 * duration */
169 if (durNap == 0)
170 durNap = 1; /* This is a very short time, make it
171 * bigger */
172
173 curpos = data + sizeof(real_ncpus); /* Point to the next
174 * available spot */
175
176 for (i = 0; i < real_ncpus; i++) { /* Move 'em all out */
177 (void) copyout((char *) &durNap, curpos, 8); /* Copy out the time
178 * since last clear */
179 (void) copyout((char *) &cpu_data_ptr[i]->cpu_hwIntCnt, curpos + 8, 256 * sizeof(uint32_t)); /* Copy out interrupt
180 * data for this
181 * processor */
182 curpos = curpos + (256 * sizeof(uint32_t) + 8); /* Point to next out put
183 * slot */
184 }
185 rval = 1;
186 (void) ml_set_interrupts_enabled(FALSE);
187 break;
188
189 case dgPowerStat:
190 {
191 uint32_t c2l = 0, c2h = 0, c3l = 0, c3h = 0, c6l = 0, c6h = 0, c7l = 0, c7h = 0;
192 uint32_t pkg_unit_l = 0, pkg_unit_h = 0, pkg_ecl = 0, pkg_ech = 0;
193
194 pkg_energy_statistics_t pkes;
195 core_energy_stat_t cest;
196
197 bzero(&pkes, sizeof(pkes));
198 bzero(&cest, sizeof(cest));
199
200 pkes.pkes_version = 1ULL;
201 rdmsr_carefully(MSR_IA32_PKG_C2_RESIDENCY, &c2l, &c2h);
202 rdmsr_carefully(MSR_IA32_PKG_C3_RESIDENCY, &c3l, &c3h);
203 rdmsr_carefully(MSR_IA32_PKG_C6_RESIDENCY, &c6l, &c6h);
204 rdmsr_carefully(MSR_IA32_PKG_C7_RESIDENCY, &c7l, &c7h);
205
206 pkes.pkg_cres[0][0] = ((uint64_t)c2h << 32) | c2l;
207 pkes.pkg_cres[0][1] = ((uint64_t)c3h << 32) | c3l;
208 pkes.pkg_cres[0][2] = ((uint64_t)c6h << 32) | c6l;
209 pkes.pkg_cres[0][3] = ((uint64_t)c7h << 32) | c7l;
210
211 uint64_t c8r = ~0ULL, c9r = ~0ULL, c10r = ~0ULL;
212
213 rdmsr64_carefully(MSR_IA32_PKG_C8_RESIDENCY, &c8r);
214 rdmsr64_carefully(MSR_IA32_PKG_C9_RESIDENCY, &c9r);
215 rdmsr64_carefully(MSR_IA32_PKG_C10_RESIDENCY, &c10r);
216
217 pkes.pkg_cres[0][4] = c8r;
218 pkes.pkg_cres[0][5] = c9r;
219 pkes.pkg_cres[0][6] = c10r;
220
221 pkes.ddr_energy = ~0ULL;
222 rdmsr64_carefully(MSR_IA32_DDR_ENERGY_STATUS, &pkes.ddr_energy);
223 pkes.llc_flushed_cycles = ~0ULL;
224 rdmsr64_carefully(MSR_IA32_LLC_FLUSHED_RESIDENCY_TIMER, &pkes.llc_flushed_cycles);
225
226 pkes.ring_ratio_instantaneous = ~0ULL;
227 rdmsr64_carefully(MSR_IA32_RING_PERF_STATUS, &pkes.ring_ratio_instantaneous);
228
229 pkes.IA_frequency_clipping_cause = ~0ULL;
230
231 uint32_t ia_perf_limits = MSR_IA32_IA_PERF_LIMIT_REASONS;
232 /* Should perhaps be a generic register map module for these
233 * registers with identical functionality that were renumbered.
234 */
235 switch (cpuid_cpufamily()) {
236 case CPUFAMILY_INTEL_SKYLAKE:
237 case CPUFAMILY_INTEL_KABYLAKE:
238 ia_perf_limits = MSR_IA32_IA_PERF_LIMIT_REASONS_SKL;
239 break;
240 default:
241 break;
242 }
243
244 rdmsr64_carefully(ia_perf_limits, &pkes.IA_frequency_clipping_cause);
245
246 pkes.GT_frequency_clipping_cause = ~0ULL;
247 rdmsr64_carefully(MSR_IA32_GT_PERF_LIMIT_REASONS, &pkes.GT_frequency_clipping_cause);
248
249 rdmsr_carefully(MSR_IA32_PKG_POWER_SKU_UNIT, &pkg_unit_l, &pkg_unit_h);
250 rdmsr_carefully(MSR_IA32_PKG_ENERGY_STATUS, &pkg_ecl, &pkg_ech);
251 pkes.pkg_power_unit = ((uint64_t)pkg_unit_h << 32) | pkg_unit_l;
252 pkes.pkg_energy = ((uint64_t)pkg_ech << 32) | pkg_ecl;
253
254 rdmsr_carefully(MSR_IA32_PP0_ENERGY_STATUS, &pkg_ecl, &pkg_ech);
255 pkes.pp0_energy = ((uint64_t)pkg_ech << 32) | pkg_ecl;
256
257 rdmsr_carefully(MSR_IA32_PP1_ENERGY_STATUS, &pkg_ecl, &pkg_ech);
258 pkes.pp1_energy = ((uint64_t)pkg_ech << 32) | pkg_ecl;
259
260 pkes.pkg_idle_exits = current_cpu_datap()->lcpu.package->package_idle_exits;
261 pkes.ncpus = real_ncpus;
262
263 (void) ml_set_interrupts_enabled(TRUE);
264
265 copyout(&pkes, regs->rsi, sizeof(pkes));
266 curpos = regs->rsi + sizeof(pkes);
267
268 mp_cpus_call(CPUMASK_ALL, ASYNC, cpu_powerstats, NULL);
269
270 for (i = 0; i < real_ncpus; i++) {
271 (void) ml_set_interrupts_enabled(FALSE);
272
273 cest.caperf = cpu_data_ptr[i]->cpu_aperf;
274 cest.cmperf = cpu_data_ptr[i]->cpu_mperf;
275 cest.ccres[0] = cpu_data_ptr[i]->cpu_c3res;
276 cest.ccres[1] = cpu_data_ptr[i]->cpu_c6res;
277 cest.ccres[2] = cpu_data_ptr[i]->cpu_c7res;
278
279 bcopy(&cpu_data_ptr[i]->cpu_rtimes[0], &cest.crtimes[0], sizeof(cest.crtimes));
280 bcopy(&cpu_data_ptr[i]->cpu_itimes[0], &cest.citimes[0], sizeof(cest.citimes));
281
282 cest.citime_total = cpu_data_ptr[i]->cpu_itime_total;
283 cest.crtime_total = cpu_data_ptr[i]->cpu_rtime_total;
284 cest.cpu_idle_exits = cpu_data_ptr[i]->cpu_idle_exits;
285 #if MONOTONIC
286 cest.cpu_insns = cpu_data_ptr[i]->cpu_monotonic.mtc_counts[MT_CORE_INSTRS];
287 cest.cpu_ucc = cpu_data_ptr[i]->cpu_monotonic.mtc_counts[MT_CORE_CYCLES];
288 cest.cpu_urc = cpu_data_ptr[i]->cpu_monotonic.mtc_counts[MT_CORE_REFCYCLES];
289 #endif /* MONOTONIC */
290 #if DIAG_ALL_PMCS
291 bcopy(&cpu_data_ptr[i]->cpu_gpmcs[0], &cest.gpmcs[0], sizeof(cest.gpmcs));
292 #endif /* DIAG_ALL_PMCS */
293 (void) ml_set_interrupts_enabled(TRUE);
294
295 copyout(&cest, curpos, sizeof(cest));
296 curpos += sizeof(cest);
297 }
298 rval = 1;
299 (void) ml_set_interrupts_enabled(FALSE);
300 }
301 break;
302 case dgEnaPMC:
303 {
304 boolean_t enable = TRUE;
305 uint32_t cpuinfo[4];
306 /* Require architectural PMC v2 or higher, corresponding to
307 * Merom+, or equivalent virtualised facility.
308 */
309 do_cpuid(0xA, &cpuinfo[0]);
310 if ((cpuinfo[0] & 0xFF) >= 2) {
311 mp_cpus_call(CPUMASK_ALL, ASYNC, cpu_pmc_control, &enable);
312 diag_pmc_enabled = TRUE;
313 }
314 rval = 1;
315 }
316 break;
317 #if DEVELOPMENT || DEBUG
318 case dgGzallocTest:
319 {
320 (void) ml_set_interrupts_enabled(TRUE);
321 if (diagflag) {
322 unsigned *ptr = (unsigned *)kalloc(1024);
323 kfree(ptr, 1024);
324 *ptr = 0x42;
325 }
326 (void) ml_set_interrupts_enabled(FALSE);
327 }
328 break;
329 #endif
330
331 #if DEVELOPMENT || DEBUG
332 case dgPermCheck:
333 {
334 (void) ml_set_interrupts_enabled(TRUE);
335 if (diagflag)
336 rval = pmap_permissions_verify(kernel_pmap, kernel_map, 0, ~0ULL);
337 (void) ml_set_interrupts_enabled(FALSE);
338 }
339 break;
340 #endif /* DEVELOPMENT || DEBUG */
341 default: /* Handle invalid ones */
342 rval = 0; /* Return an exception */
343 }
344
345 regs->rax = rval;
346
347 assert(ml_get_interrupts_enabled() == FALSE);
348 return rval;
349 }
350
351 void cpu_powerstats(__unused void *arg) {
352 cpu_data_t *cdp = current_cpu_datap();
353 __unused int cnum = cdp->cpu_number;
354 uint32_t cl = 0, ch = 0, mpl = 0, mph = 0, apl = 0, aph = 0;
355
356 rdmsr_carefully(MSR_IA32_MPERF, &mpl, &mph);
357 rdmsr_carefully(MSR_IA32_APERF, &apl, &aph);
358
359 cdp->cpu_mperf = ((uint64_t)mph << 32) | mpl;
360 cdp->cpu_aperf = ((uint64_t)aph << 32) | apl;
361
362 uint64_t ctime = mach_absolute_time();
363 cdp->cpu_rtime_total += ctime - cdp->cpu_ixtime;
364 cdp->cpu_ixtime = ctime;
365
366 rdmsr_carefully(MSR_IA32_CORE_C3_RESIDENCY, &cl, &ch);
367 cdp->cpu_c3res = ((uint64_t)ch << 32) | cl;
368
369 rdmsr_carefully(MSR_IA32_CORE_C6_RESIDENCY, &cl, &ch);
370 cdp->cpu_c6res = ((uint64_t)ch << 32) | cl;
371
372 rdmsr_carefully(MSR_IA32_CORE_C7_RESIDENCY, &cl, &ch);
373 cdp->cpu_c7res = ((uint64_t)ch << 32) | cl;
374
375 if (diag_pmc_enabled) {
376 #if MONOTONIC
377 mt_update_fixed_counts();
378 #else /* MONOTONIC */
379 uint64_t insns = read_pmc(FIXED_PMC0);
380 uint64_t ucc = read_pmc(FIXED_PMC1);
381 uint64_t urc = read_pmc(FIXED_PMC2);
382 #endif /* !MONOTONIC */
383 #if DIAG_ALL_PMCS
384 int i;
385
386 for (i = 0; i < 4; i++) {
387 cdp->cpu_gpmcs[i] = read_pmc(i);
388 }
389 #endif /* DIAG_ALL_PMCS */
390 #if !MONOTONIC
391 cdp->cpu_cur_insns = insns;
392 cdp->cpu_cur_ucc = ucc;
393 cdp->cpu_cur_urc = urc;
394 #endif /* !MONOTONIC */
395 }
396 }
397
398 void cpu_pmc_control(void *enablep) {
399 #if !MONOTONIC
400 boolean_t enable = *(boolean_t *)enablep;
401 cpu_data_t *cdp = current_cpu_datap();
402
403 if (enable) {
404 wrmsr64(0x38F, 0x70000000FULL);
405 wrmsr64(0x38D, 0x333);
406 set_cr4(get_cr4() | CR4_PCE);
407
408 } else {
409 wrmsr64(0x38F, 0);
410 wrmsr64(0x38D, 0);
411 set_cr4((get_cr4() & ~CR4_PCE));
412 }
413 cdp->cpu_fixed_pmcs_enabled = enable;
414 #else /* !MONOTONIC */
415 #pragma unused(enablep)
416 #endif /* MONOTONIC */
417 }