X-Git-Url: https://git.saurik.com/apple/xnu.git/blobdiff_plain/55e303ae13a4cf49d70f2294092726f2fffb9ef2..b0d623f7f2ae71ed96e60569f61f9a9a27016e80:/osfmk/i386/commpage/commpage.h diff --git a/osfmk/i386/commpage/commpage.h b/osfmk/i386/commpage/commpage.h index a7c8d80c4..013ca246e 100644 --- a/osfmk/i386/commpage/commpage.h +++ b/osfmk/i386/commpage/commpage.h @@ -1,16 +1,19 @@ /* - * Copyright (c) 2003 Apple Computer, Inc. All rights reserved. + * Copyright (c) 2003-2009 Apple Inc. All rights reserved. * - * @APPLE_LICENSE_HEADER_START@ - * - * Copyright (c) 1999-2003 Apple Computer, Inc. All Rights Reserved. + * @APPLE_OSREFERENCE_LICENSE_HEADER_START@ * * This file contains Original Code and/or Modifications of Original Code * as defined in and that are subject to the Apple Public Source License * Version 2.0 (the 'License'). You may not use this file except in - * compliance with the License. Please obtain a copy of the License at - * http://www.opensource.apple.com/apsl/ and read it before using this - * file. + * compliance with the License. The rights granted to you under the License + * may not be used to create, or enable the creation or redistribution of, + * unlawful or unlicensed copies of an Apple operating system, or to + * circumvent, violate, or enable the circumvention or violation of, any + * terms of an Apple operating system software license agreement. + * + * Please obtain a copy of the License at + * http://www.opensource.apple.com/apsl/ and read it before using this file. * * The Original Code and all software distributed under the License are * distributed on an 'AS IS' basis, WITHOUT WARRANTY OF ANY KIND, EITHER @@ -20,7 +23,7 @@ * Please see the License for the specific language governing rights and * limitations under the License. * - * @APPLE_LICENSE_HEADER_END@ + * @APPLE_OSREFERENCE_LICENSE_HEADER_END@ */ #ifndef _I386_COMMPAGE_H @@ -28,24 +31,95 @@ #ifndef __ASSEMBLER__ #include +#include +#include +#include #endif /* __ASSEMBLER__ */ +/* When trying to acquire a spinlock or mutex, we will spin in + * user mode for awhile, before entering the kernel to relinquish. + * MP_SPIN_TRIES is the initial value of _COMM_PAGE_SPIN_COUNT. + * The idea is that _COMM_PAGE_SPIN_COUNT will be adjusted up or + * down as the machine is plugged in/out, etc. + * At present spinlocks do not use _COMM_PAGE_SPIN_COUNT. + * They use MP_SPIN_TRIES directly. + */ +#define MP_SPIN_TRIES 1000 + + +/* The following macro is used to generate the 64-bit commpage address for a given + * routine, based on its 32-bit address. This is used in the kernel to compile + * the 64-bit commpage. Since the kernel can be a 32-bit object, cpu_capabilities.h + * only defines the 32-bit address. + */ +#define _COMM_PAGE_32_TO_64( ADDRESS ) ( ADDRESS + _COMM_PAGE64_START_ADDRESS - _COMM_PAGE32_START_ADDRESS ) + + #ifdef __ASSEMBLER__ -#include + +#define COMMPAGE_DESCRIPTOR_NAME(label) _commpage_ ## label + +#if defined (__i386__) + +#define COMMPAGE_DESCRIPTOR_FIELD_POINTER .long +#define COMMPAGE_DESCRIPTOR_REFERENCE(label) \ + .long COMMPAGE_DESCRIPTOR_NAME(label) + +#elif defined (__x86_64__) + +#define COMMPAGE_DESCRIPTOR_FIELD_POINTER .quad +#define COMMPAGE_DESCRIPTOR_REFERENCE(label) \ + .quad COMMPAGE_DESCRIPTOR_NAME(label) + +#else +#error unsupported architecture +#endif + +#define COMMPAGE_FUNCTION_START(label,codetype,alignment) \ +.text ;\ +.code ## codetype ;\ +.align alignment, 0x90 ;\ +L ## label ## : #define COMMPAGE_DESCRIPTOR(label,address,must,cant) \ -L ## label ## _end: ;\ -.const_data ;\ -L ## label ## _size = L ## label ## _end - L ## label ;\ -.private_extern _commpage_ ## label ;\ -_commpage_ ## label ## : ;\ - .long L ## label ;\ - .long L ## label ## _size ;\ - .long address ;\ - .long must ;\ - .long cant ;\ +L ## label ## _end: ;\ +.set L ## label ## _size, L ## label ## _end - L ## label ;\ +.const_data ;\ +.private_extern COMMPAGE_DESCRIPTOR_NAME(label) ;\ +COMMPAGE_DESCRIPTOR_NAME(label) ## : ;\ + COMMPAGE_DESCRIPTOR_FIELD_POINTER L ## label ;\ + .long L ## label ## _size ;\ + .long address ;\ + .long must ;\ + .long cant ;\ .text + +/* COMMPAGE_CALL(target,from,start) + * + * This macro compiles a relative near call to one + * commpage routine from another. + * The assembler cannot handle this directly because the code + * is not being assembled at the address at which it will execute. + * The alternative to this macro would be to use an + * indirect call, which is slower because the target of an + * indirect branch is poorly predicted. + * The macro arguments are: + * target = the commpage routine we are calling + * from = the commpage routine we are in now + * start = the label at the start of the code for this func + * This is admitedly ugly and fragile. Is there a better way? + */ +#define COMMPAGE_CALL(target,from,start) \ + COMMPAGE_CALL_INTERNAL(target,from,start,__LINE__) + +#define COMMPAGE_CALL_INTERNAL(target,from,start,unique) \ + .byte 0xe8 ;\ +.set UNIQUEID(unique), L ## start - . + target - from - 4 ;\ + .long UNIQUEID(unique) + +#define UNIQUEID(name) L ## name + #else /* __ASSEMBLER__ */ /* Each potential commpage routine is described by one of these. @@ -54,17 +128,41 @@ _commpage_ ## label ## : ;\ */ typedef struct commpage_descriptor { - void *code_address; // address of code - long code_length; // length in bytes - long commpage_address; // put at this address (_COMM_PAGE_BCOPY etc) - long musthave; // _cpu_capability bits we must have - long canthave; // _cpu_capability bits we can't have + void *code_address; // address of code + uint32_t code_length; // length in bytes + uint32_t commpage_address; // put at this address (_COMM_PAGE_BCOPY etc) + uint32_t musthave; // _cpu_capability bits we must have + uint32_t canthave; // _cpu_capability bits we can't have } commpage_descriptor; -extern char *commPagePtr; // virt address of commpage in kernel map +/* Warning: following structure must match the layout of the commpage. */ +/* This is the data starting at _COMM_PAGE_TIME_DATA_START, ie for nanotime() and gettimeofday() */ + +typedef volatile struct commpage_time_data { + uint64_t nt_tsc_base; // _COMM_PAGE_NT_TSC_BASE + uint32_t nt_scale; // _COMM_PAGE_NT_SCALE + uint32_t nt_shift; // _COMM_PAGE_NT_SHIFT + uint64_t nt_ns_base; // _COMM_PAGE_NT_NS_BASE + uint32_t nt_generation; // _COMM_PAGE_NT_GENERATION + uint32_t gtod_generation; // _COMM_PAGE_GTOD_GENERATION + uint64_t gtod_ns_base; // _COMM_PAGE_GTOD_NS_BASE + uint64_t gtod_sec_base; // _COMM_PAGE_GTOD_SEC_BASE +} commpage_time_data; + + +extern char *commPagePtr32; // virt address of 32-bit commpage in kernel map +extern char *commPagePtr64; // ...and of 64-bit commpage + +extern void commpage_set_timestamp(uint64_t abstime, uint64_t secs); +extern void commpage_disable_timestamp( void ); +extern void commpage_set_nanotime(uint64_t tsc_base, uint64_t ns_base, uint32_t scale, uint32_t shift); +extern void commpage_set_memory_pressure(unsigned int pressure); +extern void commpage_set_spin_count(unsigned int count); +extern void commpage_sched_gen_inc(void); -extern void commpage_set_timestamp(uint64_t tbr,uint32_t secs,uint32_t usecs,uint32_t ticks_per_sec); +extern uint32_t commpage_is_in_pfz32(uint32_t); +extern uint32_t commpage_is_in_pfz64(addr64_t); #endif /* __ASSEMBLER__ */