/*
- * Copyright (c) 2003 Apple Computer, Inc. All rights reserved.
+ * Copyright (c) 2003-2009 Apple Inc. All rights reserved.
*
- * @APPLE_LICENSE_HEADER_START@
+ * @APPLE_OSREFERENCE_LICENSE_HEADER_START@
*
- * The contents of this file constitute Original Code as defined in and
- * are subject to the Apple Public Source License Version 1.1 (the
- * "License"). You may not use this file except in compliance with the
- * License. Please obtain a copy of the License at
- * http://www.apple.com/publicsource and read it before using this file.
+ * This file contains Original Code and/or Modifications of Original Code
+ * as defined in and that are subject to the Apple Public Source License
+ * Version 2.0 (the 'License'). You may not use this file except in
+ * compliance with the License. The rights granted to you under the License
+ * may not be used to create, or enable the creation or redistribution of,
+ * unlawful or unlicensed copies of an Apple operating system, or to
+ * circumvent, violate, or enable the circumvention or violation of, any
+ * terms of an Apple operating system software license agreement.
*
- * This Original Code and all software distributed under the License are
- * distributed on an "AS IS" basis, WITHOUT WARRANTY OF ANY KIND, EITHER
+ * Please obtain a copy of the License at
+ * http://www.opensource.apple.com/apsl/ and read it before using this file.
+ *
+ * The Original Code and all software distributed under the License are
+ * distributed on an 'AS IS' basis, WITHOUT WARRANTY OF ANY KIND, EITHER
* EXPRESS OR IMPLIED, AND APPLE HEREBY DISCLAIMS ALL SUCH WARRANTIES,
* INCLUDING WITHOUT LIMITATION, ANY WARRANTIES OF MERCHANTABILITY,
- * FITNESS FOR A PARTICULAR PURPOSE OR NON-INFRINGEMENT. Please see the
- * License for the specific language governing rights and limitations
- * under the License.
+ * FITNESS FOR A PARTICULAR PURPOSE, QUIET ENJOYMENT OR NON-INFRINGEMENT.
+ * Please see the License for the specific language governing rights and
+ * limitations under the License.
*
- * @APPLE_LICENSE_HEADER_END@
+ * @APPLE_OSREFERENCE_LICENSE_HEADER_END@
*/
#ifndef _I386_COMMPAGE_H
#ifndef __ASSEMBLER__
#include <stdint.h>
+#include <mach/boolean.h>
+#include <mach/vm_types.h>
+#include <machine/cpu_capabilities.h>
#endif /* __ASSEMBLER__ */
+/* When trying to acquire a spinlock or mutex, we will spin in
+ * user mode for awhile, before entering the kernel to relinquish.
+ * MP_SPIN_TRIES is the initial value of _COMM_PAGE_SPIN_COUNT.
+ * The idea is that _COMM_PAGE_SPIN_COUNT will be adjusted up or
+ * down as the machine is plugged in/out, etc.
+ * At present spinlocks do not use _COMM_PAGE_SPIN_COUNT.
+ * They use MP_SPIN_TRIES directly.
+ */
+#define MP_SPIN_TRIES 1000
+
+
+/* The following macro is used to generate the 64-bit commpage address for a given
+ * routine, based on its 32-bit address. This is used in the kernel to compile
+ * the 64-bit commpage. Since the kernel can be a 32-bit object, cpu_capabilities.h
+ * only defines the 32-bit address.
+ */
+#define _COMM_PAGE_32_TO_64( ADDRESS ) ( ADDRESS + _COMM_PAGE64_START_ADDRESS - _COMM_PAGE32_START_ADDRESS )
+
+
#ifdef __ASSEMBLER__
-#include <machine/asm.h>
+
+#define COMMPAGE_DESCRIPTOR_NAME(label) _commpage_ ## label
+
+#if defined (__i386__)
+
+#define COMMPAGE_DESCRIPTOR_FIELD_POINTER .long
+#define COMMPAGE_DESCRIPTOR_REFERENCE(label) \
+ .long COMMPAGE_DESCRIPTOR_NAME(label)
+
+#elif defined (__x86_64__)
+
+#define COMMPAGE_DESCRIPTOR_FIELD_POINTER .quad
+#define COMMPAGE_DESCRIPTOR_REFERENCE(label) \
+ .quad COMMPAGE_DESCRIPTOR_NAME(label)
+
+#else
+#error unsupported architecture
+#endif
+
+#define COMMPAGE_FUNCTION_START(label,codetype,alignment) \
+.text ;\
+.code ## codetype ;\
+.align alignment, 0x90 ;\
+L ## label ## :
#define COMMPAGE_DESCRIPTOR(label,address,must,cant) \
-L ## label ## _end: ;\
-.const_data ;\
-L ## label ## _size = L ## label ## _end - L ## label ;\
-.private_extern _commpage_ ## label ;\
-_commpage_ ## label ## : ;\
- .long L ## label ;\
- .long L ## label ## _size ;\
- .long address ;\
- .long must ;\
- .long cant ;\
+L ## label ## _end: ;\
+.set L ## label ## _size, L ## label ## _end - L ## label ;\
+.const_data ;\
+.private_extern COMMPAGE_DESCRIPTOR_NAME(label) ;\
+COMMPAGE_DESCRIPTOR_NAME(label) ## : ;\
+ COMMPAGE_DESCRIPTOR_FIELD_POINTER L ## label ;\
+ .long L ## label ## _size ;\
+ .long address ;\
+ .long must ;\
+ .long cant ;\
.text
+
+/* COMMPAGE_CALL(target,from,start)
+ *
+ * This macro compiles a relative near call to one
+ * commpage routine from another.
+ * The assembler cannot handle this directly because the code
+ * is not being assembled at the address at which it will execute.
+ * The alternative to this macro would be to use an
+ * indirect call, which is slower because the target of an
+ * indirect branch is poorly predicted.
+ * The macro arguments are:
+ * target = the commpage routine we are calling
+ * from = the commpage routine we are in now
+ * start = the label at the start of the code for this func
+ * This is admitedly ugly and fragile. Is there a better way?
+ */
+#define COMMPAGE_CALL(target,from,start) \
+ COMMPAGE_CALL_INTERNAL(target,from,start,__LINE__)
+
+#define COMMPAGE_CALL_INTERNAL(target,from,start,unique) \
+ .byte 0xe8 ;\
+.set UNIQUEID(unique), L ## start - . + target - from - 4 ;\
+ .long UNIQUEID(unique)
+
+#define UNIQUEID(name) L ## name
+
#else /* __ASSEMBLER__ */
/* Each potential commpage routine is described by one of these.
*/
typedef struct commpage_descriptor {
- void *code_address; // address of code
- long code_length; // length in bytes
- long commpage_address; // put at this address (_COMM_PAGE_BCOPY etc)
- long musthave; // _cpu_capability bits we must have
- long canthave; // _cpu_capability bits we can't have
+ void *code_address; // address of code
+ uint32_t code_length; // length in bytes
+ uint32_t commpage_address; // put at this address (_COMM_PAGE_BCOPY etc)
+ uint32_t musthave; // _cpu_capability bits we must have
+ uint32_t canthave; // _cpu_capability bits we can't have
} commpage_descriptor;
-extern char *commPagePtr; // virt address of commpage in kernel map
+/* Warning: following structure must match the layout of the commpage. */
+/* This is the data starting at _COMM_PAGE_TIME_DATA_START, ie for nanotime() and gettimeofday() */
+
+typedef volatile struct commpage_time_data {
+ uint64_t nt_tsc_base; // _COMM_PAGE_NT_TSC_BASE
+ uint32_t nt_scale; // _COMM_PAGE_NT_SCALE
+ uint32_t nt_shift; // _COMM_PAGE_NT_SHIFT
+ uint64_t nt_ns_base; // _COMM_PAGE_NT_NS_BASE
+ uint32_t nt_generation; // _COMM_PAGE_NT_GENERATION
+ uint32_t gtod_generation; // _COMM_PAGE_GTOD_GENERATION
+ uint64_t gtod_ns_base; // _COMM_PAGE_GTOD_NS_BASE
+ uint64_t gtod_sec_base; // _COMM_PAGE_GTOD_SEC_BASE
+} commpage_time_data;
+
+
+extern char *commPagePtr32; // virt address of 32-bit commpage in kernel map
+extern char *commPagePtr64; // ...and of 64-bit commpage
-extern void commpage_set_timestamp(uint64_t tbr,uint32_t secs,uint32_t usecs,uint32_t ticks_per_sec);
+extern void commpage_set_timestamp(uint64_t abstime, uint64_t secs);
+extern void commpage_disable_timestamp( void );
+extern void commpage_set_nanotime(uint64_t tsc_base, uint64_t ns_base, uint32_t scale, uint32_t shift);
+extern void commpage_set_memory_pressure(unsigned int pressure);
+extern void commpage_set_spin_count(unsigned int count);
+extern void commpage_sched_gen_inc(void);
+extern void commpage_update_active_cpus(void);
-typedef struct {
- uint64_t nt_base_tsc;
- uint64_t nt_base_ns;
- uint32_t nt_scale;
- uint32_t nt_shift;
- uint64_t nt_check_tsc;
-} commpage_nanotime_t;
-extern void commpage_set_nanotime(commpage_nanotime_t *new_nanotime);
+extern uint32_t commpage_is_in_pfz32(uint32_t);
+extern uint32_t commpage_is_in_pfz64(addr64_t);
#endif /* __ASSEMBLER__ */