* Author: Bill Angell, Apple
* Date: 10/auht-five
*
- * Random diagnostics
+ * Random diagnostics, augmented Derek Kumar 2011
*
- * Try to keep the x86 selectors in-sync with the ppc selectors.
*
*/
#include <i386/pmCPU.h>
#include <i386/tsc.h>
#include <mach/i386/syscall_sw.h>
+#include <kern/kalloc.h>
+#include <sys/kdebug.h>
+#include <i386/machine_cpu.h>
+#include <i386/misc_protos.h>
+#include <i386/cpuid.h>
-extern uint64_t lastNapClear;
+#define PERMIT_PERMCHECK (0)
diagWork dgWork;
-uint64_t lastNapClear = 0ULL;
uint64_t lastRuptClear = 0ULL;
+boolean_t diag_pmc_enabled = FALSE;
+void cpu_powerstats(void *);
+
+typedef struct {
+ uint64_t caperf;
+ uint64_t cmperf;
+ uint64_t ccres[6];
+ uint64_t crtimes[CPU_RTIME_BINS];
+ uint64_t citimes[CPU_ITIME_BINS];
+ uint64_t crtime_total;
+ uint64_t citime_total;
+ uint64_t cpu_idle_exits;
+ uint64_t cpu_insns;
+ uint64_t cpu_ucc;
+ uint64_t cpu_urc;
+#if DIAG_ALL_PMCS
+ uint64_t gpmcs[4];
+#endif /* DIAG_ALL_PMCS */
+} core_energy_stat_t;
+
+typedef struct {
+ uint64_t pkes_version;
+ uint64_t pkg_cres[2][7];
+ uint64_t pkg_power_unit;
+ uint64_t pkg_energy;
+ uint64_t pp0_energy;
+ uint64_t pp1_energy;
+ uint64_t ddr_energy;
+ uint64_t llc_flushed_cycles;
+ uint64_t ring_ratio_instantaneous;
+ uint64_t IA_frequency_clipping_cause;
+ uint64_t GT_frequency_clipping_cause;
+ uint64_t pkg_idle_exits;
+ uint64_t pkg_rtimes[CPU_RTIME_BINS];
+ uint64_t pkg_itimes[CPU_ITIME_BINS];
+ uint64_t mbus_delay_time;
+ uint64_t mint_delay_time;
+ uint32_t ncpus;
+ core_energy_stat_t cest[];
+} pkg_energy_statistics_t;
int
diagCall64(x86_saved_state_t * state)
{
- uint64_t curpos, i, j;
- uint64_t selector, data;
- uint64_t currNap, durNap;
+ uint64_t curpos, i, j;
+ uint64_t selector, data;
+ uint64_t currNap, durNap;
x86_saved_state64_t *regs;
+ boolean_t diagflag;
+ uint32_t rval = 0;
assert(is_saved_state64(state));
regs = saved_state64(state);
- if (!(dgWork.dgFlags & enaDiagSCs))
- return 0; /* If not enabled, cause an exception */
-
+ diagflag = ((dgWork.dgFlags & enaDiagSCs) != 0);
selector = regs->rdi;
switch (selector) { /* Select the routine */
case dgRuptStat: /* Suck Interruption statistics */
+ (void) ml_set_interrupts_enabled(TRUE);
data = regs->rsi; /* Get the number of processors */
if (data == 0) { /* If no location is specified for data, clear all
}
lastRuptClear = mach_absolute_time(); /* Get the time of clear */
- return 1; /* Normal return */
+ rval = 1; /* Normal return */
+ (void) ml_set_interrupts_enabled(FALSE);
+ break;
}
(void) copyout((char *) &real_ncpus, data, sizeof(real_ncpus)); /* Copy out number of
* processors */
-
currNap = mach_absolute_time(); /* Get the time now */
durNap = currNap - lastRuptClear; /* Get the last interval
* duration */
curpos = curpos + (256 * sizeof(uint32_t) + 8); /* Point to next out put
* slot */
}
+ rval = 1;
+ (void) ml_set_interrupts_enabled(FALSE);
break;
- default: /* Handle invalid ones */
- return 0; /* Return an exception */
+ case dgPowerStat:
+ {
+ uint32_t c2l = 0, c2h = 0, c3l = 0, c3h = 0, c6l = 0, c6h = 0, c7l = 0, c7h = 0;
+ uint32_t pkg_unit_l = 0, pkg_unit_h = 0, pkg_ecl = 0, pkg_ech = 0;
+
+ pkg_energy_statistics_t pkes;
+ core_energy_stat_t cest;
+
+ bzero(&pkes, sizeof(pkes));
+ bzero(&cest, sizeof(cest));
+
+ pkes.pkes_version = 1ULL;
+ rdmsr_carefully(MSR_IA32_PKG_C2_RESIDENCY, &c2l, &c2h);
+ rdmsr_carefully(MSR_IA32_PKG_C3_RESIDENCY, &c3l, &c3h);
+ rdmsr_carefully(MSR_IA32_PKG_C6_RESIDENCY, &c6l, &c6h);
+ rdmsr_carefully(MSR_IA32_PKG_C7_RESIDENCY, &c7l, &c7h);
+
+ pkes.pkg_cres[0][0] = ((uint64_t)c2h << 32) | c2l;
+ pkes.pkg_cres[0][1] = ((uint64_t)c3h << 32) | c3l;
+ pkes.pkg_cres[0][2] = ((uint64_t)c6h << 32) | c6l;
+ pkes.pkg_cres[0][3] = ((uint64_t)c7h << 32) | c7l;
+
+ uint64_t c8r = ~0ULL, c9r = ~0ULL, c10r = ~0ULL;
+
+ rdmsr64_carefully(MSR_IA32_PKG_C8_RESIDENCY, &c8r);
+ rdmsr64_carefully(MSR_IA32_PKG_C9_RESIDENCY, &c9r);
+ rdmsr64_carefully(MSR_IA32_PKG_C10_RESIDENCY, &c10r);
+
+ pkes.pkg_cres[0][4] = c8r;
+ pkes.pkg_cres[0][5] = c9r;
+ pkes.pkg_cres[0][6] = c10r;
+
+ pkes.ddr_energy = ~0ULL;
+ rdmsr64_carefully(MSR_IA32_DDR_ENERGY_STATUS, &pkes.ddr_energy);
+ pkes.llc_flushed_cycles = ~0ULL;
+ rdmsr64_carefully(MSR_IA32_LLC_FLUSHED_RESIDENCY_TIMER, &pkes.llc_flushed_cycles);
+
+ pkes.ring_ratio_instantaneous = ~0ULL;
+ rdmsr64_carefully(MSR_IA32_RING_PERF_STATUS, &pkes.ring_ratio_instantaneous);
+
+ pkes.IA_frequency_clipping_cause = ~0ULL;
+
+ uint32_t ia_perf_limits = MSR_IA32_IA_PERF_LIMIT_REASONS;
+ /* Should perhaps be a generic register map module for these
+ * registers with identical functionality that were renumbered.
+ */
+ switch (cpuid_cpufamily()) {
+ case CPUFAMILY_INTEL_SKYLAKE:
+ ia_perf_limits = MSR_IA32_IA_PERF_LIMIT_REASONS_SKL;
+ break;
+ default:
+ break;
+ }
+
+ rdmsr64_carefully(ia_perf_limits, &pkes.IA_frequency_clipping_cause);
+
+ pkes.GT_frequency_clipping_cause = ~0ULL;
+ rdmsr64_carefully(MSR_IA32_GT_PERF_LIMIT_REASONS, &pkes.GT_frequency_clipping_cause);
+
+ rdmsr_carefully(MSR_IA32_PKG_POWER_SKU_UNIT, &pkg_unit_l, &pkg_unit_h);
+ rdmsr_carefully(MSR_IA32_PKG_ENERGY_STATUS, &pkg_ecl, &pkg_ech);
+ pkes.pkg_power_unit = ((uint64_t)pkg_unit_h << 32) | pkg_unit_l;
+ pkes.pkg_energy = ((uint64_t)pkg_ech << 32) | pkg_ecl;
+
+ rdmsr_carefully(MSR_IA32_PP0_ENERGY_STATUS, &pkg_ecl, &pkg_ech);
+ pkes.pp0_energy = ((uint64_t)pkg_ech << 32) | pkg_ecl;
+
+ rdmsr_carefully(MSR_IA32_PP1_ENERGY_STATUS, &pkg_ecl, &pkg_ech);
+ pkes.pp1_energy = ((uint64_t)pkg_ech << 32) | pkg_ecl;
+
+ pkes.pkg_idle_exits = current_cpu_datap()->lcpu.package->package_idle_exits;
+ pkes.ncpus = real_ncpus;
+
+ (void) ml_set_interrupts_enabled(TRUE);
+
+ copyout(&pkes, regs->rsi, sizeof(pkes));
+ curpos = regs->rsi + sizeof(pkes);
+
+ mp_cpus_call(CPUMASK_ALL, ASYNC, cpu_powerstats, NULL);
+
+ for (i = 0; i < real_ncpus; i++) {
+ (void) ml_set_interrupts_enabled(FALSE);
+
+ cest.caperf = cpu_data_ptr[i]->cpu_aperf;
+ cest.cmperf = cpu_data_ptr[i]->cpu_mperf;
+ cest.ccres[0] = cpu_data_ptr[i]->cpu_c3res;
+ cest.ccres[1] = cpu_data_ptr[i]->cpu_c6res;
+ cest.ccres[2] = cpu_data_ptr[i]->cpu_c7res;
+
+ bcopy(&cpu_data_ptr[i]->cpu_rtimes[0], &cest.crtimes[0], sizeof(cest.crtimes));
+ bcopy(&cpu_data_ptr[i]->cpu_itimes[0], &cest.citimes[0], sizeof(cest.citimes));
+ cest.citime_total = cpu_data_ptr[i]->cpu_itime_total;
+ cest.crtime_total = cpu_data_ptr[i]->cpu_rtime_total;
+ cest.cpu_idle_exits = cpu_data_ptr[i]->cpu_idle_exits;
+ cest.cpu_insns = cpu_data_ptr[i]->cpu_cur_insns;
+ cest.cpu_ucc = cpu_data_ptr[i]->cpu_cur_ucc;
+ cest.cpu_urc = cpu_data_ptr[i]->cpu_cur_urc;
+#if DIAG_ALL_PMCS
+ bcopy(&cpu_data_ptr[i]->cpu_gpmcs[0], &cest.gpmcs[0], sizeof(cest.gpmcs));
+#endif /* DIAG_ALL_PMCS */
+ (void) ml_set_interrupts_enabled(TRUE);
+
+ copyout(&cest, curpos, sizeof(cest));
+ curpos += sizeof(cest);
+ }
+ rval = 1;
+ (void) ml_set_interrupts_enabled(FALSE);
+ }
+ break;
+ case dgEnaPMC:
+ {
+ boolean_t enable = TRUE;
+ uint32_t cpuinfo[4];
+ /* Require architectural PMC v2 or higher, corresponding to
+ * Merom+, or equivalent virtualised facility.
+ */
+ do_cpuid(0xA, &cpuinfo[0]);
+ if ((cpuinfo[0] & 0xFF) >= 2) {
+ mp_cpus_call(CPUMASK_ALL, ASYNC, cpu_pmc_control, &enable);
+ diag_pmc_enabled = TRUE;
+ }
+ rval = 1;
+ }
+ break;
+#if DEVELOPMENT || DEBUG
+ case dgGzallocTest:
+ {
+ (void) ml_set_interrupts_enabled(TRUE);
+ if (diagflag) {
+ unsigned *ptr = (unsigned *)kalloc(1024);
+ kfree(ptr, 1024);
+ *ptr = 0x42;
+ }
+ (void) ml_set_interrupts_enabled(FALSE);
}
+ break;
+#endif
+
+#if DEVELOPMENT || DEBUG
+ case dgPermCheck:
+ {
+ (void) ml_set_interrupts_enabled(TRUE);
+ if (diagflag)
+ rval = pmap_permissions_verify(kernel_pmap, kernel_map, 0, ~0ULL);
+ (void) ml_set_interrupts_enabled(FALSE);
+ }
+ break;
+#endif /* DEVELOPMENT || DEBUG */
+ default: /* Handle invalid ones */
+ rval = 0; /* Return an exception */
+ }
+
+ regs->rax = rval;
- return 1; /* Normal non-ast check return */
+ assert(ml_get_interrupts_enabled() == FALSE);
+ return rval;
}
+void cpu_powerstats(__unused void *arg) {
+ cpu_data_t *cdp = current_cpu_datap();
+ __unused int cnum = cdp->cpu_number;
+ uint32_t cl = 0, ch = 0, mpl = 0, mph = 0, apl = 0, aph = 0;
-int
-diagCall(x86_saved_state_t * state)
-{
- uint32_t stk, curpos, i, j;
- uint32_t selector, data;
- int err;
- uint64_t currNap, durNap;
- x86_saved_state32_t *regs;
-
- assert(is_saved_state32(state));
- regs = saved_state32(state);
-
- if (!(dgWork.dgFlags & enaDiagSCs))
- return 0; /* If not enabled, cause an exception */
-
- stk = regs->uesp; /* Point to the stack */
- err = copyin((user_addr_t) (stk + 4), (char *) &selector, sizeof(uint32_t)); /* Get the selector */
- if (err) {
- return 0; /* Failed to fetch stack */
- }
- switch (selector) { /* Select the routine */
- case dgRuptStat: /* Suck Interruption statistics */
+ rdmsr_carefully(MSR_IA32_MPERF, &mpl, &mph);
+ rdmsr_carefully(MSR_IA32_APERF, &apl, &aph);
- err = copyin((user_addr_t) (stk + 8), (char *) &data, sizeof(uint32_t)); /* Get the selector */
+ cdp->cpu_mperf = ((uint64_t)mph << 32) | mpl;
+ cdp->cpu_aperf = ((uint64_t)aph << 32) | apl;
- if (data == 0) {/* If number of processors is 0, clear all
- * counts */
- for (i = 0; i < real_ncpus; i++) { /* Cycle through
- * processors */
- for (j = 0; j < 256; j++)
- cpu_data_ptr[i]->cpu_hwIntCnt[j] = 0;
- }
+ uint64_t ctime = mach_absolute_time();
+ cdp->cpu_rtime_total += ctime - cdp->cpu_ixtime;
+ cdp->cpu_ixtime = ctime;
- lastRuptClear = mach_absolute_time(); /* Get the time of clear */
- return 1; /* Normal return */
- }
+ rdmsr_carefully(MSR_IA32_CORE_C3_RESIDENCY, &cl, &ch);
+ cdp->cpu_c3res = ((uint64_t)ch << 32) | cl;
- (void) copyout((char *) &real_ncpus, data, sizeof(real_ncpus)); /* Copy out number of
- * processors */
+ rdmsr_carefully(MSR_IA32_CORE_C6_RESIDENCY, &cl, &ch);
+ cdp->cpu_c6res = ((uint64_t)ch << 32) | cl;
- currNap = mach_absolute_time(); /* Get the time now */
- durNap = currNap - lastRuptClear; /* Get the last interval
- * duration */
- if (durNap == 0)
- durNap = 1; /* This is a very short time, make it
- * bigger */
+ rdmsr_carefully(MSR_IA32_CORE_C7_RESIDENCY, &cl, &ch);
+ cdp->cpu_c7res = ((uint64_t)ch << 32) | cl;
- curpos = data + sizeof(real_ncpus); /* Point to the next
- * available spot */
+ if (diag_pmc_enabled) {
+ uint64_t insns = read_pmc(FIXED_PMC0);
+ uint64_t ucc = read_pmc(FIXED_PMC1);
+ uint64_t urc = read_pmc(FIXED_PMC2);
+#if DIAG_ALL_PMCS
+ int i;
- for (i = 0; i < real_ncpus; i++) { /* Move 'em all out */
- (void) copyout((char *) &durNap, curpos, 8); /* Copy out the time
- * since last clear */
- (void) copyout((char *) &cpu_data_ptr[i]->cpu_hwIntCnt, curpos + 8, 256 * sizeof(uint32_t)); /* Copy out interrupt
- * data for this
- * processor */
- curpos = curpos + (256 * sizeof(uint32_t) + 8); /* Point to next out put
- * slot */
+ for (i = 0; i < 4; i++) {
+ cdp->cpu_gpmcs[i] = read_pmc(i);
}
+#endif /* DIAG_ALL_PMCS */
+ cdp->cpu_cur_insns = insns;
+ cdp->cpu_cur_ucc = ucc;
+ cdp->cpu_cur_urc = urc;
+ }
+}
- break;
+void cpu_pmc_control(void *enablep) {
+ boolean_t enable = *(boolean_t *)enablep;
+ cpu_data_t *cdp = current_cpu_datap();
- default: /* Handle invalid ones */
- return 0; /* Return an exception */
+ if (enable) {
+ wrmsr64(0x38F, 0x70000000FULL);
+ wrmsr64(0x38D, 0x333);
+ set_cr4(get_cr4() | CR4_PCE);
+ } else {
+ wrmsr64(0x38F, 0);
+ wrmsr64(0x38D, 0);
+ set_cr4((get_cr4() & ~CR4_PCE));
}
-
- return 1; /* Normal non-ast check return */
+ cdp->cpu_fixed_pmcs_enabled = enable;
}