X-Git-Url: https://git.saurik.com/apple/xnu.git/blobdiff_plain/1c79356b52d46aa6b508fb032f5ae709b1f2897b..c6bf4f310a33a9262d455ea4d3f0630b1255e3fe:/bsd/kern/kern_mib.c diff --git a/bsd/kern/kern_mib.c b/bsd/kern/kern_mib.c index d6c43effe..667b17a7d 100644 --- a/bsd/kern/kern_mib.c +++ b/bsd/kern/kern_mib.c @@ -1,23 +1,29 @@ /* - * Copyright (c) 2000 Apple Computer, Inc. All rights reserved. + * Copyright (c) 2000-2007 Apple Inc. All rights reserved. * - * @APPLE_LICENSE_HEADER_START@ - * - * The contents of this file constitute Original Code as defined in and - * are subject to the Apple Public Source License Version 1.1 (the - * "License"). You may not use this file except in compliance with the - * License. Please obtain a copy of the License at - * http://www.apple.com/publicsource and read it before using this file. - * - * This Original Code and all software distributed under the License are - * distributed on an "AS IS" basis, WITHOUT WARRANTY OF ANY KIND, EITHER + * @APPLE_OSREFERENCE_LICENSE_HEADER_START@ + * + * This file contains Original Code and/or Modifications of Original Code + * as defined in and that are subject to the Apple Public Source License + * Version 2.0 (the 'License'). You may not use this file except in + * compliance with the License. The rights granted to you under the License + * may not be used to create, or enable the creation or redistribution of, + * unlawful or unlicensed copies of an Apple operating system, or to + * circumvent, violate, or enable the circumvention or violation of, any + * terms of an Apple operating system software license agreement. + * + * Please obtain a copy of the License at + * http://www.opensource.apple.com/apsl/ and read it before using this file. + * + * The Original Code and all software distributed under the License are + * distributed on an 'AS IS' basis, WITHOUT WARRANTY OF ANY KIND, EITHER * EXPRESS OR IMPLIED, AND APPLE HEREBY DISCLAIMS ALL SUCH WARRANTIES, * INCLUDING WITHOUT LIMITATION, ANY WARRANTIES OF MERCHANTABILITY, - * FITNESS FOR A PARTICULAR PURPOSE OR NON-INFRINGEMENT. Please see the - * License for the specific language governing rights and limitations - * under the License. - * - * @APPLE_LICENSE_HEADER_END@ + * FITNESS FOR A PARTICULAR PURPOSE, QUIET ENJOYMENT OR NON-INFRINGEMENT. + * Please see the License for the specific language governing rights and + * limitations under the License. + * + * @APPLE_OSREFERENCE_LICENSE_HEADER_END@ */ /*- * Copyright (c) 1982, 1986, 1989, 1993 @@ -64,29 +70,658 @@ #include #include #include -#include +#include #include #if defined(SMP) #include #endif -SYSCTL_NODE(, 0, sysctl, CTLFLAG_RW, 0, - "Sysctl internal magic"); -SYSCTL_NODE(, CTL_KERN, kern, CTLFLAG_RW, 0, - "High kernel, proc, limits &c"); -SYSCTL_NODE(, CTL_VM, vm, CTLFLAG_RW, 0, - "Virtual memory"); -SYSCTL_NODE(, CTL_VFS, vfs, CTLFLAG_RW, 0, - "File system"); -SYSCTL_NODE(, CTL_NET, net, CTLFLAG_RW, 0, - "Network, (see socket.h)"); -SYSCTL_NODE(, CTL_DEBUG, debug, CTLFLAG_RW, 0, - "Debugging"); -SYSCTL_NODE(, CTL_HW, hw, CTLFLAG_RW, 0, - "hardware"); -SYSCTL_NODE(, CTL_MACHDEP, machdep, CTLFLAG_RW, 0, - "machine dependent"); -SYSCTL_NODE(, CTL_USER, user, CTLFLAG_RW, 0, - "user-level"); +#include /* XXX prune includes */ +#include +#include +#include +#include +#include +#include +#include +#include +#include +#include +#include +#include +#include +#include +#include +#include +#include +#include +#include +#include +#include +#include +#include +#include + +extern vm_map_t bsd_pageable_map; + +#include +#include + +#include +#include + +#include +#include + +#include /* for host_info() */ + +#if defined(__i386__) || defined(__x86_64__) +#include /* for cpuid_info() */ +#endif + +#if defined(__arm__) || defined(__arm64__) +#include /* for cpuid_info() & cache_info() */ +#endif + + +#ifndef MAX +#define MAX(a, b) (a >= b ? a : b) +#endif + +/* XXX This should be in a BSD accessible Mach header, but isn't. */ +extern unsigned int vm_page_wire_count; + +static int cputype, cpusubtype, cputhreadtype, cpufamily, cpu64bit; +static uint64_t cacheconfig[10], cachesize[10]; +static int packages; + +static char * osenvironment; +static uint32_t osenvironment_size = 0; +static uint32_t ephemeral_storage = 0; +static uint32_t use_recovery_securityd = 0; + +static struct { + uint32_t ephemeral_storage:1; + uint32_t use_recovery_securityd:1; +} property_existence = {0, 0}; + +SYSCTL_NODE(, 0, sysctl, CTLFLAG_RW | CTLFLAG_LOCKED, 0, + "Sysctl internal magic"); +SYSCTL_NODE(, CTL_KERN, kern, CTLFLAG_RW | CTLFLAG_LOCKED, 0, + "High kernel, proc, limits &c"); +SYSCTL_NODE(, CTL_VM, vm, CTLFLAG_RW | CTLFLAG_LOCKED, 0, + "Virtual memory"); +SYSCTL_NODE(, CTL_VFS, vfs, CTLFLAG_RW | CTLFLAG_LOCKED, 0, + "File system"); +SYSCTL_NODE(, CTL_NET, net, CTLFLAG_RW | CTLFLAG_LOCKED, 0, + "Network, (see socket.h)"); +SYSCTL_NODE(, CTL_DEBUG, debug, CTLFLAG_RW | CTLFLAG_LOCKED, 0, + "Debugging"); +SYSCTL_NODE(, CTL_HW, hw, CTLFLAG_RW | CTLFLAG_LOCKED, 0, + "hardware"); +SYSCTL_NODE(, CTL_MACHDEP, machdep, CTLFLAG_RW | CTLFLAG_LOCKED, 0, + "machine dependent"); +SYSCTL_NODE(, CTL_USER, user, CTLFLAG_RW | CTLFLAG_LOCKED, 0, + "user-level"); + +SYSCTL_NODE(_kern, OID_AUTO, bridge, CTLFLAG_RW | CTLFLAG_LOCKED, 0, + "bridge"); + +#define SYSCTL_RETURN(r, x) SYSCTL_OUT(r, &x, sizeof(x)) + +/****************************************************************************** + * hw.* MIB + */ + +#define CTLHW_RETQUAD (1U << 31) +#define CTLHW_LOCAL (1U << 30) + +#define HW_LOCAL_CPUTHREADTYPE (1 | CTLHW_LOCAL) +#define HW_LOCAL_PHYSICALCPU (2 | CTLHW_LOCAL) +#define HW_LOCAL_PHYSICALCPUMAX (3 | CTLHW_LOCAL) +#define HW_LOCAL_LOGICALCPU (4 | CTLHW_LOCAL) +#define HW_LOCAL_LOGICALCPUMAX (5 | CTLHW_LOCAL) + + +/* + * Supporting some variables requires us to do "real" work. We + * gather some of that here. + */ +static int +sysctl_hw_generic(__unused struct sysctl_oid *oidp, __unused void *arg1, + int arg2, struct sysctl_req *req) +{ + char dummy[65]; + int epochTemp; + ml_cpu_info_t cpu_info; + int val, doquad; + long long qval; + host_basic_info_data_t hinfo; + kern_return_t kret; + mach_msg_type_number_t count = HOST_BASIC_INFO_COUNT; + + /* + * Test and mask off the 'return quad' flag. + * Note that only some things here support it. + */ + doquad = arg2 & CTLHW_RETQUAD; + arg2 &= ~CTLHW_RETQUAD; + + ml_cpu_get_info(&cpu_info); + +#define BSD_HOST 1 + kret = host_info((host_t)BSD_HOST, HOST_BASIC_INFO, (host_info_t)&hinfo, &count); + + /* + * Handle various OIDs. + * + * OIDs that can return int or quad set val and qval and then break. + * Errors and int-only values return inline. + */ + switch (arg2) { + case HW_NCPU: + if (kret == KERN_SUCCESS) { + return SYSCTL_RETURN(req, hinfo.max_cpus); + } else { + return EINVAL; + } + case HW_AVAILCPU: + if (kret == KERN_SUCCESS) { + return SYSCTL_RETURN(req, hinfo.avail_cpus); + } else { + return EINVAL; + } + case HW_LOCAL_PHYSICALCPU: + if (kret == KERN_SUCCESS) { + return SYSCTL_RETURN(req, hinfo.physical_cpu); + } else { + return EINVAL; + } + case HW_LOCAL_PHYSICALCPUMAX: + if (kret == KERN_SUCCESS) { + return SYSCTL_RETURN(req, hinfo.physical_cpu_max); + } else { + return EINVAL; + } + case HW_LOCAL_LOGICALCPU: + if (kret == KERN_SUCCESS) { + return SYSCTL_RETURN(req, hinfo.logical_cpu); + } else { + return EINVAL; + } + case HW_LOCAL_LOGICALCPUMAX: + if (kret == KERN_SUCCESS) { + return SYSCTL_RETURN(req, hinfo.logical_cpu_max); + } else { + return EINVAL; + } + case HW_PAGESIZE: + { + vm_map_t map = get_task_map(current_task()); + val = vm_map_page_size(map); + qval = (long long)val; + break; + } + case HW_CACHELINE: + val = cpu_info.cache_line_size; + qval = (long long)val; + break; + case HW_L1ICACHESIZE: + val = cpu_info.l1_icache_size; + qval = (long long)val; + break; + case HW_L1DCACHESIZE: + val = cpu_info.l1_dcache_size; + qval = (long long)val; + break; + case HW_L2CACHESIZE: + if (cpu_info.l2_cache_size == 0xFFFFFFFF) { + return EINVAL; + } + val = cpu_info.l2_cache_size; + qval = (long long)val; + break; + case HW_L3CACHESIZE: + if (cpu_info.l3_cache_size == 0xFFFFFFFF) { + return EINVAL; + } + val = cpu_info.l3_cache_size; + qval = (long long)val; + break; + + /* + * Deprecated variables. We still support these for + * backwards compatibility purposes only. + */ + case HW_MACHINE: + bzero(dummy, sizeof(dummy)); + if (!PEGetMachineName(dummy, 64)) { + return EINVAL; + } + dummy[64] = 0; + return SYSCTL_OUT(req, dummy, strlen(dummy) + 1); + case HW_MODEL: + bzero(dummy, sizeof(dummy)); + if (!PEGetModelName(dummy, 64)) { + return EINVAL; + } + dummy[64] = 0; + return SYSCTL_OUT(req, dummy, strlen(dummy) + 1); + case HW_USERMEM: + { + int usermem = mem_size - vm_page_wire_count * page_size; + + return SYSCTL_RETURN(req, usermem); + } + case HW_EPOCH: + epochTemp = PEGetPlatformEpoch(); + if (epochTemp == -1) { + return EINVAL; + } + return SYSCTL_RETURN(req, epochTemp); + case HW_VECTORUNIT: { + int vector = cpu_info.vector_unit == 0? 0 : 1; + return SYSCTL_RETURN(req, vector); + } + case HW_L2SETTINGS: + if (cpu_info.l2_cache_size == 0xFFFFFFFF) { + return EINVAL; + } + return SYSCTL_RETURN(req, cpu_info.l2_settings); + case HW_L3SETTINGS: + if (cpu_info.l3_cache_size == 0xFFFFFFFF) { + return EINVAL; + } + return SYSCTL_RETURN(req, cpu_info.l3_settings); + default: + return ENOTSUP; + } + /* + * Callers may come to us with either int or quad buffers. + */ + if (doquad) { + return SYSCTL_RETURN(req, qval); + } + return SYSCTL_RETURN(req, val); +} + +/* hw.pagesize and hw.tbfrequency are expected as 64 bit values */ +static int +sysctl_pagesize +(__unused struct sysctl_oid *oidp, __unused void *arg1, __unused int arg2, struct sysctl_req *req) +{ + vm_map_t map = get_task_map(current_task()); + long long l = vm_map_page_size(map); + return sysctl_io_number(req, l, sizeof(l), NULL, NULL); +} + +static int +sysctl_pagesize32 +(__unused struct sysctl_oid *oidp, __unused void *arg1, __unused int arg2, struct sysctl_req *req) +{ + long long l; +#if __arm64__ + l = (long long) (1 << page_shift_user32); +#else /* __arm64__ */ + l = (long long) PAGE_SIZE; +#endif /* __arm64__ */ + return sysctl_io_number(req, l, sizeof(l), NULL, NULL); +} + +static int +sysctl_tbfrequency +(__unused struct sysctl_oid *oidp, __unused void *arg1, __unused int arg2, struct sysctl_req *req) +{ + long long l = gPEClockFrequencyInfo.timebase_frequency_hz; + return sysctl_io_number(req, l, sizeof(l), NULL, NULL); +} + +/* + * Create sysctl entries coming from device tree. + * + * Entries from device tree are loaded here because DTLookupEntry() only works before + * PE_init_iokit(). Doing this also avoids the extern-C hackery to access these entries + * from IORegistry (which requires C++). + */ +void +sysctl_load_devicetree_entries(void) +{ + DTEntry chosen; + void *value; + unsigned int size; + + if (kSuccess != DTLookupEntry(0, "/chosen", &chosen)) { + return; + } + + /* load osenvironment */ + if (kSuccess == DTGetProperty(chosen, "osenvironment", (void **) &value, &size)) { + MALLOC(osenvironment, char *, size, M_TEMP, M_WAITOK); + if (osenvironment) { + memcpy(osenvironment, value, size); + osenvironment_size = size; + } + } + + /* load ephemeral_storage */ + if (kSuccess == DTGetProperty(chosen, "ephemeral-storage", (void **) &value, &size)) { + if (size == sizeof(uint32_t)) { + ephemeral_storage = *(uint32_t *)value; + property_existence.ephemeral_storage = 1; + } + } + + /* load use_recovery_securityd */ + if (kSuccess == DTGetProperty(chosen, "use-recovery-securityd", (void **) &value, &size)) { + if (size == sizeof(uint32_t)) { + use_recovery_securityd = *(uint32_t *)value; + property_existence.use_recovery_securityd = 1; + } + } +} + +static int +sysctl_osenvironment +(__unused struct sysctl_oid *oidp, __unused void *arg1, __unused int arg2, struct sysctl_req *req) +{ + if (osenvironment_size > 0) { + return SYSCTL_OUT(req, osenvironment, osenvironment_size); + } else { + return EINVAL; + } +} + +static int +sysctl_ephemeral_storage +(__unused struct sysctl_oid *oidp, __unused void *arg1, __unused int arg2, struct sysctl_req *req) +{ + if (property_existence.ephemeral_storage) { + return SYSCTL_OUT(req, &ephemeral_storage, sizeof(ephemeral_storage)); + } else { + return EINVAL; + } +} + +static int +sysctl_use_recovery_securityd +(__unused struct sysctl_oid *oidp, __unused void *arg1, __unused int arg2, struct sysctl_req *req) +{ + if (property_existence.use_recovery_securityd) { + return SYSCTL_OUT(req, &use_recovery_securityd, sizeof(use_recovery_securityd)); + } else { + return EINVAL; + } +} + +/* + * hw.* MIB variables. + */ +SYSCTL_PROC(_hw, HW_NCPU, ncpu, CTLTYPE_INT | CTLFLAG_RD | CTLFLAG_KERN | CTLFLAG_LOCKED, 0, HW_NCPU, sysctl_hw_generic, "I", ""); +SYSCTL_PROC(_hw, HW_AVAILCPU, activecpu, CTLTYPE_INT | CTLFLAG_RD | CTLFLAG_KERN | CTLFLAG_LOCKED, 0, HW_AVAILCPU, sysctl_hw_generic, "I", ""); +SYSCTL_PROC(_hw, OID_AUTO, physicalcpu, CTLTYPE_INT | CTLFLAG_RD | CTLFLAG_KERN | CTLFLAG_LOCKED, 0, HW_LOCAL_PHYSICALCPU, sysctl_hw_generic, "I", ""); +SYSCTL_PROC(_hw, OID_AUTO, physicalcpu_max, CTLTYPE_INT | CTLFLAG_RD | CTLFLAG_KERN | CTLFLAG_LOCKED, 0, HW_LOCAL_PHYSICALCPUMAX, sysctl_hw_generic, "I", ""); +SYSCTL_PROC(_hw, OID_AUTO, logicalcpu, CTLTYPE_INT | CTLFLAG_RD | CTLFLAG_KERN | CTLFLAG_LOCKED, 0, HW_LOCAL_LOGICALCPU, sysctl_hw_generic, "I", ""); +SYSCTL_PROC(_hw, OID_AUTO, logicalcpu_max, CTLTYPE_INT | CTLFLAG_RD | CTLFLAG_KERN | CTLFLAG_LOCKED, 0, HW_LOCAL_LOGICALCPUMAX, sysctl_hw_generic, "I", ""); +SYSCTL_INT(_hw, HW_BYTEORDER, byteorder, CTLFLAG_RD | CTLFLAG_KERN | CTLFLAG_LOCKED, (int *)NULL, BYTE_ORDER, ""); +SYSCTL_INT(_hw, OID_AUTO, cputype, CTLFLAG_RD | CTLFLAG_KERN | CTLFLAG_LOCKED, &cputype, 0, ""); +SYSCTL_INT(_hw, OID_AUTO, cpusubtype, CTLFLAG_RD | CTLFLAG_KERN | CTLFLAG_LOCKED, &cpusubtype, 0, ""); +SYSCTL_INT(_hw, OID_AUTO, cpu64bit_capable, CTLFLAG_RD | CTLFLAG_KERN | CTLFLAG_LOCKED, &cpu64bit, 0, ""); +SYSCTL_INT(_hw, OID_AUTO, cpufamily, CTLFLAG_RD | CTLFLAG_KERN | CTLFLAG_LOCKED, &cpufamily, 0, ""); +SYSCTL_OPAQUE(_hw, OID_AUTO, cacheconfig, CTLFLAG_RD | CTLFLAG_LOCKED, &cacheconfig, sizeof(cacheconfig), "Q", ""); +SYSCTL_OPAQUE(_hw, OID_AUTO, cachesize, CTLFLAG_RD | CTLFLAG_LOCKED, &cachesize, sizeof(cachesize), "Q", ""); +SYSCTL_PROC(_hw, OID_AUTO, pagesize, CTLTYPE_QUAD | CTLFLAG_RD | CTLFLAG_KERN | CTLFLAG_LOCKED, 0, 0, sysctl_pagesize, "Q", ""); +SYSCTL_PROC(_hw, OID_AUTO, pagesize32, CTLTYPE_QUAD | CTLFLAG_RD | CTLFLAG_KERN | CTLFLAG_LOCKED, 0, 0, sysctl_pagesize32, "Q", ""); +#if DEBUG || DEVELOPMENT || (!defined(__arm__) && !defined(__arm64__)) +SYSCTL_QUAD(_hw, OID_AUTO, busfrequency, CTLFLAG_RD | CTLFLAG_KERN | CTLFLAG_LOCKED, &gPEClockFrequencyInfo.bus_frequency_hz, ""); +SYSCTL_QUAD(_hw, OID_AUTO, busfrequency_min, CTLFLAG_RD | CTLFLAG_KERN | CTLFLAG_LOCKED, &gPEClockFrequencyInfo.bus_frequency_min_hz, ""); +SYSCTL_QUAD(_hw, OID_AUTO, busfrequency_max, CTLFLAG_RD | CTLFLAG_KERN | CTLFLAG_LOCKED, &gPEClockFrequencyInfo.bus_frequency_max_hz, ""); +SYSCTL_QUAD(_hw, OID_AUTO, cpufrequency, CTLFLAG_RD | CTLFLAG_KERN | CTLFLAG_LOCKED, &gPEClockFrequencyInfo.cpu_frequency_hz, ""); +SYSCTL_QUAD(_hw, OID_AUTO, cpufrequency_min, CTLFLAG_RD | CTLFLAG_KERN | CTLFLAG_LOCKED, &gPEClockFrequencyInfo.cpu_frequency_min_hz, ""); +SYSCTL_QUAD(_hw, OID_AUTO, cpufrequency_max, CTLFLAG_RD | CTLFLAG_KERN | CTLFLAG_LOCKED, &gPEClockFrequencyInfo.cpu_frequency_max_hz, ""); +#endif +SYSCTL_PROC(_hw, OID_AUTO, cachelinesize, CTLTYPE_QUAD | CTLFLAG_RD | CTLFLAG_KERN | CTLFLAG_LOCKED, 0, HW_CACHELINE | CTLHW_RETQUAD, sysctl_hw_generic, "Q", ""); +SYSCTL_PROC(_hw, OID_AUTO, l1icachesize, CTLTYPE_QUAD | CTLFLAG_RD | CTLFLAG_KERN | CTLFLAG_LOCKED, 0, HW_L1ICACHESIZE | CTLHW_RETQUAD, sysctl_hw_generic, "Q", ""); +SYSCTL_PROC(_hw, OID_AUTO, l1dcachesize, CTLTYPE_QUAD | CTLFLAG_RD | CTLFLAG_KERN | CTLFLAG_LOCKED, 0, HW_L1DCACHESIZE | CTLHW_RETQUAD, sysctl_hw_generic, "Q", ""); +SYSCTL_PROC(_hw, OID_AUTO, l2cachesize, CTLTYPE_QUAD | CTLFLAG_RD | CTLFLAG_KERN | CTLFLAG_LOCKED, 0, HW_L2CACHESIZE | CTLHW_RETQUAD, sysctl_hw_generic, "Q", ""); +SYSCTL_PROC(_hw, OID_AUTO, l3cachesize, CTLTYPE_QUAD | CTLFLAG_RD | CTLFLAG_KERN | CTLFLAG_LOCKED, 0, HW_L3CACHESIZE | CTLHW_RETQUAD, sysctl_hw_generic, "Q", ""); +#if (defined(__arm__) || defined(__arm64__)) && (DEBUG || DEVELOPMENT) +SYSCTL_QUAD(_hw, OID_AUTO, memfrequency, CTLFLAG_RD | CTLFLAG_KERN | CTLFLAG_LOCKED, &gPEClockFrequencyInfo.mem_frequency_hz, ""); +SYSCTL_QUAD(_hw, OID_AUTO, memfrequency_min, CTLFLAG_RD | CTLFLAG_KERN | CTLFLAG_LOCKED, &gPEClockFrequencyInfo.mem_frequency_min_hz, ""); +SYSCTL_QUAD(_hw, OID_AUTO, memfrequency_max, CTLFLAG_RD | CTLFLAG_KERN | CTLFLAG_LOCKED, &gPEClockFrequencyInfo.mem_frequency_max_hz, ""); +SYSCTL_QUAD(_hw, OID_AUTO, prffrequency, CTLFLAG_RD | CTLFLAG_KERN | CTLFLAG_LOCKED, &gPEClockFrequencyInfo.prf_frequency_hz, ""); +SYSCTL_QUAD(_hw, OID_AUTO, prffrequency_min, CTLFLAG_RD | CTLFLAG_KERN | CTLFLAG_LOCKED, &gPEClockFrequencyInfo.prf_frequency_min_hz, ""); +SYSCTL_QUAD(_hw, OID_AUTO, prffrequency_max, CTLFLAG_RD | CTLFLAG_KERN | CTLFLAG_LOCKED, &gPEClockFrequencyInfo.prf_frequency_max_hz, ""); +SYSCTL_QUAD(_hw, OID_AUTO, fixfrequency, CTLFLAG_RD | CTLFLAG_KERN | CTLFLAG_LOCKED, &gPEClockFrequencyInfo.fix_frequency_hz, ""); +#endif /* __arm__ || __arm64__ */ +SYSCTL_PROC(_hw, OID_AUTO, tbfrequency, CTLTYPE_QUAD | CTLFLAG_RD | CTLFLAG_KERN | CTLFLAG_LOCKED, 0, 0, sysctl_tbfrequency, "Q", ""); +SYSCTL_QUAD(_hw, HW_MEMSIZE, memsize, CTLFLAG_RD | CTLFLAG_KERN | CTLFLAG_LOCKED, &max_mem, ""); +SYSCTL_INT(_hw, OID_AUTO, packages, CTLFLAG_RD | CTLFLAG_KERN | CTLFLAG_LOCKED, &packages, 0, ""); +SYSCTL_PROC(_hw, OID_AUTO, osenvironment, CTLTYPE_STRING | CTLFLAG_RD | CTLFLAG_KERN | CTLFLAG_LOCKED, 0, 0, sysctl_osenvironment, "A", ""); +SYSCTL_PROC(_hw, OID_AUTO, ephemeral_storage, CTLTYPE_INT | CTLFLAG_RD | CTLFLAG_KERN | CTLFLAG_LOCKED, 0, 0, sysctl_ephemeral_storage, "I", ""); +SYSCTL_PROC(_hw, OID_AUTO, use_recovery_securityd, CTLTYPE_INT | CTLFLAG_RD | CTLFLAG_KERN | CTLFLAG_LOCKED, 0, 0, sysctl_use_recovery_securityd, "I", ""); + +/* + * Optional CPU features can register nodes below hw.optional. + * + * If the feature is not present, the node should either not be registered, + * or it should return -1. If the feature is present, the node should return + * 0. If the feature is present and its use is advised, the node should + * return 1. + */ +SYSCTL_NODE(_hw, OID_AUTO, optional, CTLFLAG_RW | CTLFLAG_LOCKED, NULL, "optional features"); + +SYSCTL_INT(_hw_optional, OID_AUTO, floatingpoint, CTLFLAG_RD | CTLFLAG_KERN | CTLFLAG_LOCKED, (int *)NULL, 1, ""); /* always set */ + +/* + * Optional device hardware features can be registered by drivers below hw.features + */ +SYSCTL_NODE(_hw, OID_AUTO, features, CTLFLAG_RD | CTLFLAG_LOCKED, NULL, "hardware features"); + +/* + * Deprecated variables. These are supported for backwards compatibility + * purposes only. The MASKED flag requests that the variables not be + * printed by sysctl(8) and similar utilities. + * + * The variables named *_compat here are int-sized versions of variables + * that are now exported as quads. The int-sized versions are normally + * looked up only by number, wheras the quad-sized versions should be + * looked up by name. + * + * The *_compat nodes are *NOT* visible within the kernel. + */ +SYSCTL_PROC(_hw, HW_PAGESIZE, pagesize_compat, CTLTYPE_INT | CTLFLAG_RD | CTLFLAG_MASKED | CTLFLAG_LOCKED, 0, HW_PAGESIZE, sysctl_hw_generic, "I", ""); +#if DEBUG || DEVELOPMENT || (!defined(__arm__) && !defined(__arm64__)) +SYSCTL_COMPAT_INT(_hw, HW_BUS_FREQ, busfrequency_compat, CTLFLAG_RD | CTLFLAG_MASKED | CTLFLAG_LOCKED, &gPEClockFrequencyInfo.bus_clock_rate_hz, 0, ""); +SYSCTL_COMPAT_INT(_hw, HW_CPU_FREQ, cpufrequency_compat, CTLFLAG_RD | CTLFLAG_MASKED | CTLFLAG_LOCKED, &gPEClockFrequencyInfo.cpu_clock_rate_hz, 0, ""); +#endif +SYSCTL_PROC(_hw, HW_CACHELINE, cachelinesize_compat, CTLTYPE_INT | CTLFLAG_RD | CTLFLAG_MASKED | CTLFLAG_LOCKED, 0, HW_CACHELINE, sysctl_hw_generic, "I", ""); +SYSCTL_PROC(_hw, HW_L1ICACHESIZE, l1icachesize_compat, CTLTYPE_INT | CTLFLAG_RD | CTLFLAG_MASKED | CTLFLAG_LOCKED, 0, HW_L1ICACHESIZE, sysctl_hw_generic, "I", ""); +SYSCTL_PROC(_hw, HW_L1DCACHESIZE, l1dcachesize_compat, CTLTYPE_INT | CTLFLAG_RD | CTLFLAG_MASKED | CTLFLAG_LOCKED, 0, HW_L1DCACHESIZE, sysctl_hw_generic, "I", ""); +SYSCTL_PROC(_hw, HW_L2CACHESIZE, l2cachesize_compat, CTLTYPE_INT | CTLFLAG_RD | CTLFLAG_MASKED | CTLFLAG_LOCKED, 0, HW_L2CACHESIZE, sysctl_hw_generic, "I", ""); +SYSCTL_PROC(_hw, HW_L3CACHESIZE, l3cachesize_compat, CTLTYPE_INT | CTLFLAG_RD | CTLFLAG_MASKED | CTLFLAG_LOCKED, 0, HW_L3CACHESIZE, sysctl_hw_generic, "I", ""); +SYSCTL_COMPAT_INT(_hw, HW_TB_FREQ, tbfrequency_compat, CTLFLAG_RD | CTLFLAG_MASKED | CTLFLAG_LOCKED, &gPEClockFrequencyInfo.timebase_frequency_hz, 0, ""); +SYSCTL_PROC(_hw, HW_MACHINE, machine, CTLTYPE_STRING | CTLFLAG_RD | CTLFLAG_MASKED | CTLFLAG_LOCKED, 0, HW_MACHINE, sysctl_hw_generic, "A", ""); +SYSCTL_PROC(_hw, HW_MODEL, model, CTLTYPE_STRING | CTLFLAG_RD | CTLFLAG_MASKED | CTLFLAG_LOCKED, 0, HW_MODEL, sysctl_hw_generic, "A", ""); +SYSCTL_COMPAT_UINT(_hw, HW_PHYSMEM, physmem, CTLFLAG_RD | CTLFLAG_MASKED | CTLFLAG_LOCKED, &mem_size, 0, ""); +SYSCTL_PROC(_hw, HW_USERMEM, usermem, CTLTYPE_INT | CTLFLAG_RD | CTLFLAG_MASKED | CTLFLAG_LOCKED, 0, HW_USERMEM, sysctl_hw_generic, "I", ""); +SYSCTL_PROC(_hw, HW_EPOCH, epoch, CTLTYPE_INT | CTLFLAG_RD | CTLFLAG_MASKED | CTLFLAG_LOCKED, 0, HW_EPOCH, sysctl_hw_generic, "I", ""); +SYSCTL_PROC(_hw, HW_VECTORUNIT, vectorunit, CTLTYPE_INT | CTLFLAG_RD | CTLFLAG_MASKED | CTLFLAG_LOCKED, 0, HW_VECTORUNIT, sysctl_hw_generic, "I", ""); +SYSCTL_PROC(_hw, HW_L2SETTINGS, l2settings, CTLTYPE_INT | CTLFLAG_RD | CTLFLAG_MASKED | CTLFLAG_LOCKED, 0, HW_L2SETTINGS, sysctl_hw_generic, "I", ""); +SYSCTL_PROC(_hw, HW_L3SETTINGS, l3settings, CTLTYPE_INT | CTLFLAG_RD | CTLFLAG_MASKED | CTLFLAG_LOCKED, 0, HW_L3SETTINGS, sysctl_hw_generic, "I", ""); +SYSCTL_INT(_hw, OID_AUTO, cputhreadtype, CTLFLAG_RD | CTLFLAG_NOAUTO | CTLFLAG_KERN | CTLFLAG_LOCKED, &cputhreadtype, 0, ""); + +#if defined(__i386__) || defined(__x86_64__) +static int +sysctl_cpu_capability +(__unused struct sysctl_oid *oidp, void *arg1, __unused int arg2, struct sysctl_req *req) +{ + uint64_t mask = (uint64_t) (uintptr_t) arg1; + boolean_t is_capable = (_get_cpu_capabilities() & mask) != 0; + + return SYSCTL_OUT(req, &is_capable, sizeof(is_capable)); +} + +SYSCTL_PROC(_hw_optional, OID_AUTO, mmx, CTLTYPE_INT | CTLFLAG_RD | CTLFLAG_KERN | CTLFLAG_LOCKED, (void *) kHasMMX, 0, sysctl_cpu_capability, "I", ""); +SYSCTL_PROC(_hw_optional, OID_AUTO, sse, CTLTYPE_INT | CTLFLAG_RD | CTLFLAG_KERN | CTLFLAG_LOCKED, (void *) kHasSSE, 0, sysctl_cpu_capability, "I", ""); +SYSCTL_PROC(_hw_optional, OID_AUTO, sse2, CTLTYPE_INT | CTLFLAG_RD | CTLFLAG_KERN | CTLFLAG_LOCKED, (void *) kHasSSE2, 0, sysctl_cpu_capability, "I", ""); +SYSCTL_PROC(_hw_optional, OID_AUTO, sse3, CTLTYPE_INT | CTLFLAG_RD | CTLFLAG_KERN | CTLFLAG_LOCKED, (void *) kHasSSE3, 0, sysctl_cpu_capability, "I", ""); +SYSCTL_PROC(_hw_optional, OID_AUTO, supplementalsse3, CTLTYPE_INT | CTLFLAG_RD | CTLFLAG_KERN | CTLFLAG_LOCKED, (void *) kHasSupplementalSSE3, 0, sysctl_cpu_capability, "I", ""); +SYSCTL_PROC(_hw_optional, OID_AUTO, sse4_1, CTLTYPE_INT | CTLFLAG_RD | CTLFLAG_KERN | CTLFLAG_LOCKED, (void *) kHasSSE4_1, 0, sysctl_cpu_capability, "I", ""); +SYSCTL_PROC(_hw_optional, OID_AUTO, sse4_2, CTLTYPE_INT | CTLFLAG_RD | CTLFLAG_KERN | CTLFLAG_LOCKED, (void *) kHasSSE4_2, 0, sysctl_cpu_capability, "I", ""); +/* "x86_64" is actually a preprocessor symbol on the x86_64 kernel, so we have to hack this */ +#undef x86_64 +SYSCTL_PROC(_hw_optional, OID_AUTO, x86_64, CTLTYPE_INT | CTLFLAG_RD | CTLFLAG_KERN | CTLFLAG_LOCKED, (void *) k64Bit, 0, sysctl_cpu_capability, "I", ""); +SYSCTL_PROC(_hw_optional, OID_AUTO, aes, CTLTYPE_INT | CTLFLAG_RD | CTLFLAG_KERN | CTLFLAG_LOCKED, (void *) kHasAES, 0, sysctl_cpu_capability, "I", ""); +SYSCTL_PROC(_hw_optional, OID_AUTO, avx1_0, CTLTYPE_INT | CTLFLAG_RD | CTLFLAG_KERN | CTLFLAG_LOCKED, (void *) kHasAVX1_0, 0, sysctl_cpu_capability, "I", ""); +SYSCTL_PROC(_hw_optional, OID_AUTO, rdrand, CTLTYPE_INT | CTLFLAG_RD | CTLFLAG_KERN | CTLFLAG_LOCKED, (void *) kHasRDRAND, 0, sysctl_cpu_capability, "I", ""); +SYSCTL_PROC(_hw_optional, OID_AUTO, f16c, CTLTYPE_INT | CTLFLAG_RD | CTLFLAG_KERN | CTLFLAG_LOCKED, (void *) kHasF16C, 0, sysctl_cpu_capability, "I", ""); +SYSCTL_PROC(_hw_optional, OID_AUTO, enfstrg, CTLTYPE_INT | CTLFLAG_RD | CTLFLAG_KERN | CTLFLAG_LOCKED, (void *) kHasENFSTRG, 0, sysctl_cpu_capability, "I", ""); +SYSCTL_PROC(_hw_optional, OID_AUTO, fma, CTLTYPE_INT | CTLFLAG_RD | CTLFLAG_KERN | CTLFLAG_LOCKED, (void *) kHasFMA, 0, sysctl_cpu_capability, "I", ""); +SYSCTL_PROC(_hw_optional, OID_AUTO, avx2_0, CTLTYPE_INT | CTLFLAG_RD | CTLFLAG_KERN | CTLFLAG_LOCKED, (void *) kHasAVX2_0, 0, sysctl_cpu_capability, "I", ""); +SYSCTL_PROC(_hw_optional, OID_AUTO, bmi1, CTLTYPE_INT | CTLFLAG_RD | CTLFLAG_KERN | CTLFLAG_LOCKED, (void *) kHasBMI1, 0, sysctl_cpu_capability, "I", ""); +SYSCTL_PROC(_hw_optional, OID_AUTO, bmi2, CTLTYPE_INT | CTLFLAG_RD | CTLFLAG_KERN | CTLFLAG_LOCKED, (void *) kHasBMI2, 0, sysctl_cpu_capability, "I", ""); +SYSCTL_PROC(_hw_optional, OID_AUTO, rtm, CTLTYPE_INT | CTLFLAG_RD | CTLFLAG_KERN | CTLFLAG_LOCKED, (void *) kHasRTM, 0, sysctl_cpu_capability, "I", ""); +SYSCTL_PROC(_hw_optional, OID_AUTO, hle, CTLTYPE_INT | CTLFLAG_RD | CTLFLAG_KERN | CTLFLAG_LOCKED, (void *) kHasHLE, 0, sysctl_cpu_capability, "I", ""); +SYSCTL_PROC(_hw_optional, OID_AUTO, adx, CTLTYPE_INT | CTLFLAG_RD | CTLFLAG_KERN | CTLFLAG_LOCKED, (void *) kHasADX, 0, sysctl_cpu_capability, "I", ""); +SYSCTL_PROC(_hw_optional, OID_AUTO, mpx, CTLTYPE_INT | CTLFLAG_RD | CTLFLAG_KERN | CTLFLAG_LOCKED, (void *) kHasMPX, 0, sysctl_cpu_capability, "I", ""); +SYSCTL_PROC(_hw_optional, OID_AUTO, sgx, CTLTYPE_INT | CTLFLAG_RD | CTLFLAG_KERN | CTLFLAG_LOCKED, (void *) kHasSGX, 0, sysctl_cpu_capability, "I", ""); +#if !defined(RC_HIDE_XNU_J137) +SYSCTL_PROC(_hw_optional, OID_AUTO, avx512f, CTLTYPE_INT | CTLFLAG_RD | CTLFLAG_KERN | CTLFLAG_LOCKED, (void *) kHasAVX512F, 0, sysctl_cpu_capability, "I", ""); +SYSCTL_PROC(_hw_optional, OID_AUTO, avx512cd, CTLTYPE_INT | CTLFLAG_RD | CTLFLAG_KERN | CTLFLAG_LOCKED, (void *) kHasAVX512CD, 0, sysctl_cpu_capability, "I", ""); +SYSCTL_PROC(_hw_optional, OID_AUTO, avx512dq, CTLTYPE_INT | CTLFLAG_RD | CTLFLAG_KERN | CTLFLAG_LOCKED, (void *) kHasAVX512DQ, 0, sysctl_cpu_capability, "I", ""); +SYSCTL_PROC(_hw_optional, OID_AUTO, avx512bw, CTLTYPE_INT | CTLFLAG_RD | CTLFLAG_KERN | CTLFLAG_LOCKED, (void *) kHasAVX512BW, 0, sysctl_cpu_capability, "I", ""); +SYSCTL_PROC(_hw_optional, OID_AUTO, avx512vl, CTLTYPE_INT | CTLFLAG_RD | CTLFLAG_KERN | CTLFLAG_LOCKED, (void *) kHasAVX512VL, 0, sysctl_cpu_capability, "I", ""); +SYSCTL_PROC(_hw_optional, OID_AUTO, avx512ifma, CTLTYPE_INT | CTLFLAG_RD | CTLFLAG_KERN | CTLFLAG_LOCKED, (void *) kHasAVX512IFMA, 0, sysctl_cpu_capability, "I", ""); +SYSCTL_PROC(_hw_optional, OID_AUTO, avx512vbmi, CTLTYPE_INT | CTLFLAG_RD | CTLFLAG_KERN | CTLFLAG_LOCKED, (void *) kHasAVX512VBMI, 0, sysctl_cpu_capability, "I", ""); +#endif /* not RC_HIDE_XNU_J137 */ +#elif defined (__arm__) || defined (__arm64__) +int watchpoint_flag = -1; +int breakpoint_flag = -1; +int gNeon = -1; +int gNeonHpfp = -1; +int gNeonFp16 = -1; +int gARMv81Atomics = 0; +int gARMv8Crc32 = 0; +int gARMv82FHM = 0; + +#if defined (__arm__) +int arm64_flag = 0; +#elif defined (__arm64__) /* end __arm__*/ +int arm64_flag = 1; +#else /* end __arm64__*/ +int arm64_flag = -1; +#endif + +SYSCTL_INT(_hw_optional, OID_AUTO, watchpoint, CTLFLAG_RD | CTLFLAG_KERN | CTLFLAG_LOCKED, &watchpoint_flag, 0, ""); +SYSCTL_INT(_hw_optional, OID_AUTO, breakpoint, CTLFLAG_RD | CTLFLAG_KERN | CTLFLAG_LOCKED, &breakpoint_flag, 0, ""); +SYSCTL_INT(_hw_optional, OID_AUTO, neon, CTLFLAG_RD | CTLFLAG_KERN | CTLFLAG_LOCKED, &gNeon, 0, ""); +SYSCTL_INT(_hw_optional, OID_AUTO, neon_hpfp, CTLFLAG_RD | CTLFLAG_KERN | CTLFLAG_LOCKED, &gNeonHpfp, 0, ""); +SYSCTL_INT(_hw_optional, OID_AUTO, neon_fp16, CTLFLAG_RD | CTLFLAG_KERN | CTLFLAG_LOCKED, &gNeonFp16, 0, ""); +SYSCTL_INT(_hw_optional, OID_AUTO, armv8_1_atomics, CTLFLAG_RD | CTLFLAG_KERN | CTLFLAG_LOCKED, &gARMv81Atomics, 0, ""); +SYSCTL_INT(_hw_optional, OID_AUTO, armv8_crc32, CTLFLAG_RD | CTLFLAG_KERN | CTLFLAG_LOCKED, &gARMv8Crc32, 0, ""); +SYSCTL_INT(_hw_optional, OID_AUTO, armv8_2_fhm, CTLFLAG_RD | CTLFLAG_KERN | CTLFLAG_LOCKED, &gARMv82FHM, 0, ""); + +/* + * Without this little ifdef dance, the preprocessor replaces "arm64" with "1", + * leaving us with a less-than-helpful sysctl.hwoptional.1. + */ +#ifdef arm64 +#undef arm64 +SYSCTL_INT(_hw_optional, OID_AUTO, arm64, CTLFLAG_RD | CTLFLAG_KERN | CTLFLAG_LOCKED, &arm64_flag, 0, ""); +#define arm64 1 +#else +SYSCTL_INT(_hw_optional, OID_AUTO, arm64, CTLFLAG_RD | CTLFLAG_KERN | CTLFLAG_LOCKED, &arm64_flag, 0, ""); +#endif + +#else +#error Unsupported arch +#endif /* !__i386__ && !__x86_64 && !__arm__ && ! __arm64__ */ + + +/****************************************************************************** + * Generic MIB initialisation. + * + * This is a hack, and should be replaced with SYSINITs + * at some point. + */ +void +sysctl_mib_init(void) +{ + cputype = cpu_type(); + cpusubtype = cpu_subtype(); + cputhreadtype = cpu_threadtype(); +#if defined(__i386__) || defined (__x86_64__) + cpu64bit = (_get_cpu_capabilities() & k64Bit) == k64Bit; +#elif defined(__arm__) || defined (__arm64__) + cpu64bit = (cputype & CPU_ARCH_ABI64) == CPU_ARCH_ABI64; +#else +#error Unsupported arch +#endif + + /* + * Populate the optional portion of the hw.* MIB. + * + * XXX This could be broken out into parts of the code + * that actually directly relate to the functions in + * question. + */ + + if (cputhreadtype != CPU_THREADTYPE_NONE) { + sysctl_register_oid(&sysctl__hw_cputhreadtype); + } + +#if defined (__i386__) || defined (__x86_64__) + /* hw.cpufamily */ + cpufamily = cpuid_cpufamily(); + + /* hw.cacheconfig */ + cacheconfig[0] = ml_cpu_cache_sharing(0); + cacheconfig[1] = ml_cpu_cache_sharing(1); + cacheconfig[2] = ml_cpu_cache_sharing(2); + cacheconfig[3] = ml_cpu_cache_sharing(3); + cacheconfig[4] = 0; + + /* hw.cachesize */ + cachesize[0] = ml_cpu_cache_size(0); + cachesize[1] = ml_cpu_cache_size(1); + cachesize[2] = ml_cpu_cache_size(2); + cachesize[3] = ml_cpu_cache_size(3); + cachesize[4] = 0; + + /* hw.packages */ + packages = roundup(ml_cpu_cache_sharing(0), cpuid_info()->thread_count) + / cpuid_info()->thread_count; + +#elif defined(__arm__) || defined(__arm64__) /* end __i386 */ + + cpufamily = cpuid_get_cpufamily(); + + watchpoint_flag = arm_debug_info()->num_watchpoint_pairs; + breakpoint_flag = arm_debug_info()->num_breakpoint_pairs; + + arm_mvfp_info_t *mvfp_info = arm_mvfp_info(); + gNeon = mvfp_info->neon; + gNeonHpfp = mvfp_info->neon_hpfp; + gNeonFp16 = mvfp_info->neon_fp16; + + cacheconfig[0] = ml_get_max_cpus(); + cacheconfig[1] = 1; + cacheconfig[2] = cache_info()->c_l2size ? 1:0; + cacheconfig[3] = 0; + cacheconfig[4] = 0; + cacheconfig[5] = 0; + cacheconfig[6] = 0; + + cachesize[0] = ml_get_machine_mem(); + cachesize[1] = cache_info()->c_dsize; /* Using the DCache */ + cachesize[2] = cache_info()->c_l2size; + cachesize[3] = 0; + cachesize[4] = 0; + + packages = 1; +#else +#error unknown architecture +#endif /* !__i386__ && !__x86_64 && !__arm__ && !__arm64__ */ +}