]> git.saurik.com Git - apple/xnu.git/blobdiff - osfmk/vm/pmap.h
xnu-2050.7.9.tar.gz
[apple/xnu.git] / osfmk / vm / pmap.h
index bf7cbed6cfe79385783a420bc2a20e3405f72d8f..26c26d6c0b2cd5af98ce5f50a2a28046105c62ab 100644 (file)
@@ -169,9 +169,7 @@ extern void                 mapping_free_prime(void); /* Primes the mapping block release list
  */
 
 extern boolean_t       pmap_next_page(ppnum_t *pnum);
-#if defined(__LP64__)
-extern boolean_t       pmap_next_page_k64(ppnum_t *pnum);
-#endif
+extern boolean_t       pmap_next_page_hi(ppnum_t *pnum);
                                                /* During VM initialization,
                                                 * return the next unused
                                                 * physical page.
@@ -189,6 +187,7 @@ extern void         pmap_virtual_space(
  *     Routines to manage the physical map data structure.
  */
 extern pmap_t          pmap_create(    /* Create a pmap_t. */
+                               ledger_t        ledger,
                                vm_map_size_t   size,
 #ifdef __i386__
                                boolean_t       is_64bit);
@@ -206,6 +205,7 @@ extern void         pmap_enter(     /* Enter a mapping */
                                vm_map_offset_t v,
                                ppnum_t         pn,
                                vm_prot_t       prot,
+                               vm_prot_t       fault_type,
                                unsigned int    flags,
                                boolean_t       wired);
 
@@ -214,6 +214,7 @@ extern kern_return_t        pmap_enter_options(
                                           vm_map_offset_t v,
                                           ppnum_t pn,
                                           vm_prot_t prot,
+                                          vm_prot_t fault_type,
                                           unsigned int flags,
                                           boolean_t wired,
                                           unsigned int options);
@@ -276,6 +277,12 @@ extern kern_return_t       (pmap_attribute_cache_sync)(  /* Flush appropriate
 extern unsigned int    (pmap_cache_attributes)(
                                ppnum_t         pn);
 
+/*
+ * Set (override) cache attributes for the specified physical page
+ */
+extern void            pmap_set_cache_attributes(
+                               ppnum_t,
+                               unsigned int);
 extern void pmap_sync_page_data_phys(ppnum_t pa);
 extern void pmap_sync_page_attributes_phys(ppnum_t pa);
 
@@ -370,24 +377,25 @@ extern kern_return_t      (pmap_attribute)(       /* Get/Set special memory
 /*
  *     Macro to be used in place of pmap_enter()
  */
-#define PMAP_ENTER(pmap, virtual_address, page, protection, flags, wired) \
+#define PMAP_ENTER(pmap, virtual_address, page, protection, fault_type, flags, wired) \
        MACRO_BEGIN                                                     \
        pmap_t          __pmap = (pmap);                                \
        vm_page_t       __page = (page);                                \
                                                                        \
        PMAP_ENTER_CHECK(__pmap, __page)                                \
-       pmap_enter(__pmap,                                      \
+       pmap_enter(__pmap,                                              \
                (virtual_address),                                      \
                __page->phys_page,                                      \
-                       (protection),                                   \
+               (protection),                                           \
+               (fault_type),                                           \
                (flags),                                                \
                (wired));                                               \
        MACRO_END
 #endif /* !PMAP_ENTER */
 
 #ifndef        PMAP_ENTER_OPTIONS
-#define PMAP_ENTER_OPTIONS(pmap, virtual_address, page, protection,    \
-                               flags, wired, options, result) \
+#define PMAP_ENTER_OPTIONS(pmap, virtual_address, page, protection, fault_type,        \
+                               flags, wired, options, result)          \
        MACRO_BEGIN                                                     \
        pmap_t          __pmap = (pmap);                                \
        vm_page_t       __page = (page);                                \
@@ -396,13 +404,41 @@ extern kern_return_t      (pmap_attribute)(       /* Get/Set special memory
        result = pmap_enter_options(__pmap,                             \
                (virtual_address),                                      \
                __page->phys_page,                                      \
-                       (protection),                                   \
+               (protection),                                           \
+               (fault_type),                                           \
                (flags),                                                \
                (wired),                                                \
-               options);                                       \
+               options);                                               \
        MACRO_END
 #endif /* !PMAP_ENTER_OPTIONS */
 
+#ifndef PMAP_SET_CACHE_ATTR
+#define PMAP_SET_CACHE_ATTR(mem, object, cache_attr, batch_pmap_op)            \
+       MACRO_BEGIN                                                             \
+               if (!batch_pmap_op) {                                           \
+                       pmap_set_cache_attributes(mem->phys_page, cache_attr);  \
+                       object->set_cache_attr = TRUE;                          \
+               }                                                               \
+       MACRO_END                                                       
+#endif /* PMAP_SET_CACHE_ATTR */
+
+#ifndef PMAP_BATCH_SET_CACHE_ATTR
+#define PMAP_BATCH_SET_CACHE_ATTR(object, user_page_list,                      \
+                                       cache_attr, num_pages, batch_pmap_op)   \
+       MACRO_BEGIN                                                             \
+               if ((batch_pmap_op)) {                                          \
+                       unsigned int __page_idx=0;                              \
+                       while (__page_idx < (num_pages)) {                      \
+                               pmap_set_cache_attributes(                      \
+                                       user_page_list[__page_idx].phys_addr,   \
+                                       (cache_attr));                          \
+                               __page_idx++;                                   \
+                       }                                                       \
+                       (object)->set_cache_attr = TRUE;                        \
+               }                                                               \
+       MACRO_END
+#endif /* PMAP_BATCH_SET_CACHE_ATTR */
+
 #define PMAP_ENTER_CHECK(pmap, page)                                   \
 {                                                                      \
        if ((pmap) != kernel_pmap) {                                    \
@@ -455,17 +491,21 @@ extern void               (pmap_pageable)(
 extern uint64_t pmap_nesting_size_min;
 extern uint64_t pmap_nesting_size_max;
 
-extern kern_return_t pmap_nest(pmap_t grand,
-                              pmap_t subord,
-                              addr64_t vstart,
-                              addr64_t nstart,
-                              uint64_t size);
-extern kern_return_t pmap_unnest(pmap_t grand,
-                                addr64_t vaddr,
-                                uint64_t size);
+extern kern_return_t pmap_nest(pmap_t,
+                              pmap_t,
+                              addr64_t,
+                              addr64_t,
+                              uint64_t);
+extern kern_return_t pmap_unnest(pmap_t,
+                                addr64_t,
+                                uint64_t);
 extern boolean_t pmap_adjust_unnest_parameters(pmap_t, vm_map_offset_t *, vm_map_offset_t *);
 #endif /* MACH_KERNEL_PRIVATE */
 
+extern boolean_t       pmap_is_noencrypt(ppnum_t);
+extern void            pmap_set_noencrypt(ppnum_t pn);
+extern void            pmap_clear_noencrypt(ppnum_t pn);
+
 /*
  * JMM - This portion is exported to other kernel components right now,
  * but will be pulled back in the future when the needed functionality
@@ -482,14 +522,18 @@ extern pmap_t     kernel_pmap;                    /* The kernel's map */
 #define VM_MEM_NOT_CACHEABLE   0x4             /* (I) Cache Inhibit */
 #define VM_MEM_WRITE_THROUGH   0x8             /* (W) Write-Through */
 
+#define VM_WIMG_USE_DEFAULT    0x80
 #define VM_WIMG_MASK           0xFF
-#define VM_WIMG_USE_DEFAULT    0x80000000
 
 #define VM_MEM_SUPERPAGE       0x100           /* map a superpage instead of a base page */
+#define VM_MEM_STACK           0x200
 
 #define PMAP_OPTIONS_NOWAIT    0x1             /* don't block, return 
                                                 * KERN_RESOURCE_SHORTAGE 
                                                 * instead */
+#define PMAP_OPTIONS_NOENTER   0x2             /* expand pmap if needed
+                                                * but don't enter mapping
+                                                */
 
 #if    !defined(__LP64__)
 extern vm_offset_t     pmap_extract(pmap_t pmap,