2 * Copyright (c) 2005-2008 Apple Inc. All rights reserved.
4 * @APPLE_OSREFERENCE_LICENSE_HEADER_START@
6 * This file contains Original Code and/or Modifications of Original Code
7 * as defined in and that are subject to the Apple Public Source License
8 * Version 2.0 (the 'License'). You may not use this file except in
9 * compliance with the License. The rights granted to you under the License
10 * may not be used to create, or enable the creation or redistribution of,
11 * unlawful or unlicensed copies of an Apple operating system, or to
12 * circumvent, violate, or enable the circumvention or violation of, any
13 * terms of an Apple operating system software license agreement.
15 * Please obtain a copy of the License at
16 * http://www.opensource.apple.com/apsl/ and read it before using this file.
18 * The Original Code and all software distributed under the License are
19 * distributed on an 'AS IS' basis, WITHOUT WARRANTY OF ANY KIND, EITHER
20 * EXPRESS OR IMPLIED, AND APPLE HEREBY DISCLAIMS ALL SUCH WARRANTIES,
21 * INCLUDING WITHOUT LIMITATION, ANY WARRANTIES OF MERCHANTABILITY,
22 * FITNESS FOR A PARTICULAR PURPOSE, QUIET ENJOYMENT OR NON-INFRINGEMENT.
23 * Please see the License for the specific language governing rights and
24 * limitations under the License.
26 * @APPLE_OSREFERENCE_LICENSE_HEADER_END@
29 * @OSF_FREE_COPYRIGHT@
32 * @APPLE_FREE_COPYRIGHT@
36 * Author: Bill Angell, Apple
39 * Random diagnostics, augmented Derek Kumar 2011
45 #include <kern/machine.h>
46 #include <kern/processor.h>
47 #include <mach/machine.h>
48 #include <mach/processor_info.h>
49 #include <mach/mach_types.h>
50 #include <mach/boolean.h>
51 #include <kern/thread.h>
52 #include <kern/task.h>
53 #include <kern/ipc_kobject.h>
54 #include <mach/vm_param.h>
56 #include <ipc/ipc_entry.h>
57 #include <ipc/ipc_space.h>
58 #include <ipc/ipc_object.h>
59 #include <ipc/ipc_port.h>
60 #include <vm/vm_kern.h>
61 #include <vm/vm_map.h>
62 #include <vm/vm_page.h>
64 #include <pexpert/pexpert.h>
65 #include <console/video_console.h>
66 #include <i386/cpu_data.h>
67 #include <i386/Diagnostics.h>
69 #include <i386/pmCPU.h>
71 #include <mach/i386/syscall_sw.h>
72 #include <kern/kalloc.h>
73 #include <sys/kdebug.h>
74 #include <i386/machine_cpu.h>
75 #include <i386/misc_protos.h>
76 #include <i386/cpuid.h>
78 #define PERMIT_PERMCHECK (0)
81 uint64_t lastRuptClear
= 0ULL;
82 boolean_t diag_pmc_enabled
= FALSE
;
83 void cpu_powerstats(void *);
89 uint64_t crtimes
[CPU_RTIME_BINS
];
90 uint64_t citimes
[CPU_ITIME_BINS
];
91 uint64_t crtime_total
;
92 uint64_t citime_total
;
93 uint64_t cpu_idle_exits
;
99 #endif /* DIAG_ALL_PMCS */
100 } core_energy_stat_t
;
103 uint64_t pkes_version
;
104 uint64_t pkg_cres
[2][7];
105 uint64_t pkg_power_unit
;
110 uint64_t llc_flushed_cycles
;
111 uint64_t ring_ratio_instantaneous
;
112 uint64_t IA_frequency_clipping_cause
;
113 uint64_t GT_frequency_clipping_cause
;
114 uint64_t pkg_idle_exits
;
115 uint64_t pkg_rtimes
[CPU_RTIME_BINS
];
116 uint64_t pkg_itimes
[CPU_ITIME_BINS
];
117 uint64_t mbus_delay_time
;
118 uint64_t mint_delay_time
;
120 core_energy_stat_t cest
[];
121 } pkg_energy_statistics_t
;
125 diagCall64(x86_saved_state_t
* state
)
127 uint64_t curpos
, i
, j
;
128 uint64_t selector
, data
;
129 uint64_t currNap
, durNap
;
130 x86_saved_state64_t
*regs
;
134 assert(is_saved_state64(state
));
135 regs
= saved_state64(state
);
137 diagflag
= ((dgWork
.dgFlags
& enaDiagSCs
) != 0);
138 selector
= regs
->rdi
;
140 switch (selector
) { /* Select the routine */
141 case dgRuptStat
: /* Suck Interruption statistics */
142 (void) ml_set_interrupts_enabled(TRUE
);
143 data
= regs
->rsi
; /* Get the number of processors */
145 if (data
== 0) { /* If no location is specified for data, clear all
148 for (i
= 0; i
< real_ncpus
; i
++) { /* Cycle through
150 for (j
= 0; j
< 256; j
++)
151 cpu_data_ptr
[i
]->cpu_hwIntCnt
[j
] = 0;
154 lastRuptClear
= mach_absolute_time(); /* Get the time of clear */
155 rval
= 1; /* Normal return */
156 (void) ml_set_interrupts_enabled(FALSE
);
160 (void) copyout((char *) &real_ncpus
, data
, sizeof(real_ncpus
)); /* Copy out number of
162 currNap
= mach_absolute_time(); /* Get the time now */
163 durNap
= currNap
- lastRuptClear
; /* Get the last interval
166 durNap
= 1; /* This is a very short time, make it
169 curpos
= data
+ sizeof(real_ncpus
); /* Point to the next
172 for (i
= 0; i
< real_ncpus
; i
++) { /* Move 'em all out */
173 (void) copyout((char *) &durNap
, curpos
, 8); /* Copy out the time
174 * since last clear */
175 (void) copyout((char *) &cpu_data_ptr
[i
]->cpu_hwIntCnt
, curpos
+ 8, 256 * sizeof(uint32_t)); /* Copy out interrupt
178 curpos
= curpos
+ (256 * sizeof(uint32_t) + 8); /* Point to next out put
182 (void) ml_set_interrupts_enabled(FALSE
);
187 uint32_t c2l
= 0, c2h
= 0, c3l
= 0, c3h
= 0, c6l
= 0, c6h
= 0, c7l
= 0, c7h
= 0;
188 uint32_t pkg_unit_l
= 0, pkg_unit_h
= 0, pkg_ecl
= 0, pkg_ech
= 0;
190 pkg_energy_statistics_t pkes
;
191 core_energy_stat_t cest
;
193 bzero(&pkes
, sizeof(pkes
));
194 bzero(&cest
, sizeof(cest
));
196 pkes
.pkes_version
= 1ULL;
197 rdmsr_carefully(MSR_IA32_PKG_C2_RESIDENCY
, &c2l
, &c2h
);
198 rdmsr_carefully(MSR_IA32_PKG_C3_RESIDENCY
, &c3l
, &c3h
);
199 rdmsr_carefully(MSR_IA32_PKG_C6_RESIDENCY
, &c6l
, &c6h
);
200 rdmsr_carefully(MSR_IA32_PKG_C7_RESIDENCY
, &c7l
, &c7h
);
202 pkes
.pkg_cres
[0][0] = ((uint64_t)c2h
<< 32) | c2l
;
203 pkes
.pkg_cres
[0][1] = ((uint64_t)c3h
<< 32) | c3l
;
204 pkes
.pkg_cres
[0][2] = ((uint64_t)c6h
<< 32) | c6l
;
205 pkes
.pkg_cres
[0][3] = ((uint64_t)c7h
<< 32) | c7l
;
207 uint64_t c8r
= ~0ULL, c9r
= ~0ULL, c10r
= ~0ULL;
209 rdmsr64_carefully(MSR_IA32_PKG_C8_RESIDENCY
, &c8r
);
210 rdmsr64_carefully(MSR_IA32_PKG_C9_RESIDENCY
, &c9r
);
211 rdmsr64_carefully(MSR_IA32_PKG_C10_RESIDENCY
, &c10r
);
213 pkes
.pkg_cres
[0][4] = c8r
;
214 pkes
.pkg_cres
[0][5] = c9r
;
215 pkes
.pkg_cres
[0][6] = c10r
;
217 pkes
.ddr_energy
= ~0ULL;
218 rdmsr64_carefully(MSR_IA32_DDR_ENERGY_STATUS
, &pkes
.ddr_energy
);
219 pkes
.llc_flushed_cycles
= ~0ULL;
220 rdmsr64_carefully(MSR_IA32_LLC_FLUSHED_RESIDENCY_TIMER
, &pkes
.llc_flushed_cycles
);
222 pkes
.ring_ratio_instantaneous
= ~0ULL;
223 rdmsr64_carefully(MSR_IA32_RING_PERF_STATUS
, &pkes
.ring_ratio_instantaneous
);
225 pkes
.IA_frequency_clipping_cause
= ~0ULL;
227 uint32_t ia_perf_limits
= MSR_IA32_IA_PERF_LIMIT_REASONS
;
228 /* Should perhaps be a generic register map module for these
229 * registers with identical functionality that were renumbered.
231 switch (cpuid_cpufamily()) {
232 case CPUFAMILY_INTEL_SKYLAKE
:
233 ia_perf_limits
= MSR_IA32_IA_PERF_LIMIT_REASONS_SKL
;
239 rdmsr64_carefully(ia_perf_limits
, &pkes
.IA_frequency_clipping_cause
);
241 pkes
.GT_frequency_clipping_cause
= ~0ULL;
242 rdmsr64_carefully(MSR_IA32_GT_PERF_LIMIT_REASONS
, &pkes
.GT_frequency_clipping_cause
);
244 rdmsr_carefully(MSR_IA32_PKG_POWER_SKU_UNIT
, &pkg_unit_l
, &pkg_unit_h
);
245 rdmsr_carefully(MSR_IA32_PKG_ENERGY_STATUS
, &pkg_ecl
, &pkg_ech
);
246 pkes
.pkg_power_unit
= ((uint64_t)pkg_unit_h
<< 32) | pkg_unit_l
;
247 pkes
.pkg_energy
= ((uint64_t)pkg_ech
<< 32) | pkg_ecl
;
249 rdmsr_carefully(MSR_IA32_PP0_ENERGY_STATUS
, &pkg_ecl
, &pkg_ech
);
250 pkes
.pp0_energy
= ((uint64_t)pkg_ech
<< 32) | pkg_ecl
;
252 rdmsr_carefully(MSR_IA32_PP1_ENERGY_STATUS
, &pkg_ecl
, &pkg_ech
);
253 pkes
.pp1_energy
= ((uint64_t)pkg_ech
<< 32) | pkg_ecl
;
255 pkes
.pkg_idle_exits
= current_cpu_datap()->lcpu
.package
->package_idle_exits
;
256 pkes
.ncpus
= real_ncpus
;
258 (void) ml_set_interrupts_enabled(TRUE
);
260 copyout(&pkes
, regs
->rsi
, sizeof(pkes
));
261 curpos
= regs
->rsi
+ sizeof(pkes
);
263 mp_cpus_call(CPUMASK_ALL
, ASYNC
, cpu_powerstats
, NULL
);
265 for (i
= 0; i
< real_ncpus
; i
++) {
266 (void) ml_set_interrupts_enabled(FALSE
);
268 cest
.caperf
= cpu_data_ptr
[i
]->cpu_aperf
;
269 cest
.cmperf
= cpu_data_ptr
[i
]->cpu_mperf
;
270 cest
.ccres
[0] = cpu_data_ptr
[i
]->cpu_c3res
;
271 cest
.ccres
[1] = cpu_data_ptr
[i
]->cpu_c6res
;
272 cest
.ccres
[2] = cpu_data_ptr
[i
]->cpu_c7res
;
274 bcopy(&cpu_data_ptr
[i
]->cpu_rtimes
[0], &cest
.crtimes
[0], sizeof(cest
.crtimes
));
275 bcopy(&cpu_data_ptr
[i
]->cpu_itimes
[0], &cest
.citimes
[0], sizeof(cest
.citimes
));
277 cest
.citime_total
= cpu_data_ptr
[i
]->cpu_itime_total
;
278 cest
.crtime_total
= cpu_data_ptr
[i
]->cpu_rtime_total
;
279 cest
.cpu_idle_exits
= cpu_data_ptr
[i
]->cpu_idle_exits
;
280 cest
.cpu_insns
= cpu_data_ptr
[i
]->cpu_cur_insns
;
281 cest
.cpu_ucc
= cpu_data_ptr
[i
]->cpu_cur_ucc
;
282 cest
.cpu_urc
= cpu_data_ptr
[i
]->cpu_cur_urc
;
284 bcopy(&cpu_data_ptr
[i
]->cpu_gpmcs
[0], &cest
.gpmcs
[0], sizeof(cest
.gpmcs
));
285 #endif /* DIAG_ALL_PMCS */
286 (void) ml_set_interrupts_enabled(TRUE
);
288 copyout(&cest
, curpos
, sizeof(cest
));
289 curpos
+= sizeof(cest
);
292 (void) ml_set_interrupts_enabled(FALSE
);
297 boolean_t enable
= TRUE
;
299 /* Require architectural PMC v2 or higher, corresponding to
300 * Merom+, or equivalent virtualised facility.
302 do_cpuid(0xA, &cpuinfo
[0]);
303 if ((cpuinfo
[0] & 0xFF) >= 2) {
304 mp_cpus_call(CPUMASK_ALL
, ASYNC
, cpu_pmc_control
, &enable
);
305 diag_pmc_enabled
= TRUE
;
313 (void) ml_set_interrupts_enabled(TRUE
);
315 unsigned *ptr
= (unsigned *)kalloc(1024);
319 (void) ml_set_interrupts_enabled(FALSE
);
327 (void) ml_set_interrupts_enabled(TRUE
);
329 rval
= pmap_permissions_verify(kernel_pmap
, kernel_map
, 0, ~0ULL);
330 (void) ml_set_interrupts_enabled(FALSE
);
333 #endif /* PERMIT_PERMCHECK */
334 default: /* Handle invalid ones */
335 rval
= 0; /* Return an exception */
340 assert(ml_get_interrupts_enabled() == FALSE
);
344 void cpu_powerstats(__unused
void *arg
) {
345 cpu_data_t
*cdp
= current_cpu_datap();
346 __unused
int cnum
= cdp
->cpu_number
;
347 uint32_t cl
= 0, ch
= 0, mpl
= 0, mph
= 0, apl
= 0, aph
= 0;
349 rdmsr_carefully(MSR_IA32_MPERF
, &mpl
, &mph
);
350 rdmsr_carefully(MSR_IA32_APERF
, &apl
, &aph
);
352 cdp
->cpu_mperf
= ((uint64_t)mph
<< 32) | mpl
;
353 cdp
->cpu_aperf
= ((uint64_t)aph
<< 32) | apl
;
355 uint64_t ctime
= mach_absolute_time();
356 cdp
->cpu_rtime_total
+= ctime
- cdp
->cpu_ixtime
;
357 cdp
->cpu_ixtime
= ctime
;
359 rdmsr_carefully(MSR_IA32_CORE_C3_RESIDENCY
, &cl
, &ch
);
360 cdp
->cpu_c3res
= ((uint64_t)ch
<< 32) | cl
;
362 rdmsr_carefully(MSR_IA32_CORE_C6_RESIDENCY
, &cl
, &ch
);
363 cdp
->cpu_c6res
= ((uint64_t)ch
<< 32) | cl
;
365 rdmsr_carefully(MSR_IA32_CORE_C7_RESIDENCY
, &cl
, &ch
);
366 cdp
->cpu_c7res
= ((uint64_t)ch
<< 32) | cl
;
368 if (diag_pmc_enabled
) {
369 uint64_t insns
= read_pmc(FIXED_PMC0
);
370 uint64_t ucc
= read_pmc(FIXED_PMC1
);
371 uint64_t urc
= read_pmc(FIXED_PMC2
);
375 for (i
= 0; i
< 4; i
++) {
376 cdp
->cpu_gpmcs
[i
] = read_pmc(i
);
378 #endif /* DIAG_ALL_PMCS */
379 cdp
->cpu_cur_insns
= insns
;
380 cdp
->cpu_cur_ucc
= ucc
;
381 cdp
->cpu_cur_urc
= urc
;
385 void cpu_pmc_control(void *enablep
) {
386 boolean_t enable
= *(boolean_t
*)enablep
;
387 cpu_data_t
*cdp
= current_cpu_datap();
390 wrmsr64(0x38F, 0x70000000FULL
);
391 wrmsr64(0x38D, 0x333);
392 set_cr4(get_cr4() | CR4_PCE
);
397 set_cr4((get_cr4() & ~CR4_PCE
));
399 cdp
->cpu_fixed_pmcs_enabled
= enable
;