]> git.saurik.com Git - apple/xnu.git/blobdiff - osfmk/i386/commpage/commpage.h
xnu-3247.10.11.tar.gz
[apple/xnu.git] / osfmk / i386 / commpage / commpage.h
index dd63874cf2485c73420ebd707bab31a91ea7cbf1..6f2a3418d6860577dbc73274d5f179cade5d5e44 100644 (file)
@@ -1,5 +1,5 @@
 /*
- * Copyright (c) 2003-2006 Apple Computer, Inc. All rights reserved.
+ * Copyright (c) 2003-2012 Apple Inc. All rights reserved.
  *
  * @APPLE_OSREFERENCE_LICENSE_HEADER_START@
  * 
 
 #ifndef        __ASSEMBLER__
 #include <stdint.h>
+#include <mach/boolean.h>
+#include <mach/vm_types.h>
+#include <machine/cpu_capabilities.h>
 #endif /* __ASSEMBLER__ */
 
+/* When trying to acquire a spinlock or mutex, we will spin in
+ * user mode for awhile, before entering the kernel to relinquish.
+ * MP_SPIN_TRIES is the initial value of _COMM_PAGE_SPIN_COUNT.
+ * The idea is that _COMM_PAGE_SPIN_COUNT will be adjusted up or
+ * down as the machine is plugged in/out, etc.
+ * At present spinlocks do not use _COMM_PAGE_SPIN_COUNT.
+ * They use MP_SPIN_TRIES directly.
+ */
+#define        MP_SPIN_TRIES   1000
+
+
 /* The following macro is used to generate the 64-bit commpage address for a given
  * routine, based on its 32-bit address.  This is used in the kernel to compile
- * the 64-bit commpage.  Since the kernel is a 32-bit object, cpu_capabilities.h
+ * the 64-bit commpage.  Since the kernel can be a 32-bit object, cpu_capabilities.h
  * only defines the 32-bit address.
  */
 #define        _COMM_PAGE_32_TO_64( ADDRESS )  ( ADDRESS + _COMM_PAGE64_START_ADDRESS - _COMM_PAGE32_START_ADDRESS )
 
 #ifdef __ASSEMBLER__
 
-#define        COMMPAGE_DESCRIPTOR(label,address,must,cant)    \
-L ## label ## _end:                                    ;\
-.const_data                                            ;\
-L ## label ## _size = L ## label ## _end - L ## label  ;\
-.private_extern _commpage_ ## label                    ;\
-_commpage_ ## label ## :                               ;\
-    .long      L ## label                              ;\
-    .long      L ## label ## _size                     ;\
-    .long      address                                 ;\
-    .long      must                                    ;\
-    .long      cant                                    ;\
+#define COMMPAGE_DESCRIPTOR_NAME(label)  _commpage_ ## label
+
+#define COMMPAGE_DESCRIPTOR_FIELD_POINTER .quad
+#define COMMPAGE_DESCRIPTOR_REFERENCE(label) \
+       .quad COMMPAGE_DESCRIPTOR_NAME(label)
+
+#define COMMPAGE_FUNCTION_START(label,codetype,alignment) \
+.text                                                          ;\
+.code ## codetype                                              ;\
+.align alignment, 0x90                                         ;\
+L ## label ## :
+
+#define        COMMPAGE_DESCRIPTOR(label,address)                      \
+L ## label ## _end:                                            ;\
+.set L ## label ## _size, L ## label ## _end - L ## label      ;\
+.const_data                                                    ;\
+.private_extern COMMPAGE_DESCRIPTOR_NAME(label)                        ;\
+COMMPAGE_DESCRIPTOR_NAME(label) ## :                           ;\
+    COMMPAGE_DESCRIPTOR_FIELD_POINTER  L ## label              ;\
+    .long                              L ## label ## _size     ;\
+    .long                              address                 ;\
 .text
 
+
+/* COMMPAGE_CALL(target,from,start)
+ *
+ * This macro compiles a relative near call to one
+ * commpage routine from another.
+ * The assembler cannot handle this directly because the code
+ * is not being assembled at the address at which it will execute.
+ * The alternative to this macro would be to use an
+ * indirect call, which is slower because the target of an
+ * indirect branch is poorly predicted.
+ * The macro arguments are:
+ *     target = the commpage routine we are calling
+ *     from   = the commpage routine we are in now
+ *     start  = the label at the start of the code for this func
+ * This is admitedly ugly and fragile.  Is there a better way?
+ */
+#define COMMPAGE_CALL(target,from,start)                       \
+       COMMPAGE_CALL_INTERNAL(target,from,start,__LINE__)
+
+#define COMMPAGE_CALL_INTERNAL(target,from,start,unique)       \
+       .byte 0xe8                                              ;\
+.set UNIQUEID(unique), L ## start - . + target - from - 4      ;\
+       .long   UNIQUEID(unique)
+
+#define UNIQUEID(name) L ## name
+
 #else /* __ASSEMBLER__ */
 
 /* Each potential commpage routine is described by one of these.
@@ -64,23 +114,43 @@ _commpage_ ## label ## :                           ;\
  */
  
 typedef        struct  commpage_descriptor     {
-    void       *code_address;                                  // address of code
-    long       code_length;                                    // length in bytes
-    long       commpage_address;                               // put at this address (_COMM_PAGE_BCOPY etc)
-    long       musthave;                                       // _cpu_capability bits we must have
-    long       canthave;                                       // _cpu_capability bits we can't have
+    void               *code_address;                          // address of code
+    uint32_t           code_length;                            // length in bytes
+    uint32_t           commpage_address;                       // put at this address (_COMM_PAGE_BCOPY etc)
 } commpage_descriptor;
 
 
+/* Warning: following structure must match the layout of the commpage.  */
+/* This is the data starting at _COMM_PAGE_TIME_DATA_START, ie for nanotime() and gettimeofday() */
+
+typedef        volatile struct commpage_time_data      {
+       uint64_t        nt_tsc_base;                            // _COMM_PAGE_NT_TSC_BASE
+       uint32_t        nt_scale;                               // _COMM_PAGE_NT_SCALE
+       uint32_t        nt_shift;                               // _COMM_PAGE_NT_SHIFT
+       uint64_t        nt_ns_base;                             // _COMM_PAGE_NT_NS_BASE
+       uint32_t        nt_generation;                          // _COMM_PAGE_NT_GENERATION
+       uint32_t        gtod_generation;                        // _COMM_PAGE_GTOD_GENERATION
+       uint64_t        gtod_ns_base;                           // _COMM_PAGE_GTOD_NS_BASE
+       uint64_t        gtod_sec_base;                          // _COMM_PAGE_GTOD_SEC_BASE
+} commpage_time_data;
+
+
 extern char    *commPagePtr32;                         // virt address of 32-bit commpage in kernel map
 extern char    *commPagePtr64;                         // ...and of 64-bit commpage
 
-extern void    _commpage_set_timestamp(uint64_t abstime, uint64_t secs);
-#define commpage_set_timestamp(x, y, z)                _commpage_set_timestamp((x), (y))
-
+extern void    commpage_set_timestamp(uint64_t abstime, uint64_t secs);
+extern void    commpage_disable_timestamp( void );
 extern  void   commpage_set_nanotime(uint64_t tsc_base, uint64_t ns_base, uint32_t scale, uint32_t shift);
+extern void    commpage_set_memory_pressure(unsigned int  pressure);
+extern void    commpage_set_spin_count(unsigned int  count);
+extern void    commpage_sched_gen_inc(void);
+extern void    commpage_update_active_cpus(void);
+extern void    commpage_update_mach_approximate_time(uint64_t abstime);
+extern void    commpage_update_kdebug_enable(void);
+extern void    commpage_update_atm_diagnostic_config(uint32_t);
 
-#include <kern/page_decrypt.h>
+extern uint32_t        commpage_is_in_pfz32(uint32_t);
+extern uint32_t        commpage_is_in_pfz64(addr64_t);
 
 #endif /* __ASSEMBLER__ */