* External memory management interface control functions.
*/
-#ifdef MACH_BSD
-/* THIS code should be removed when the component merge is completed */
-extern int vnode_pager_workaround;
-#endif
-
#include <advisory_pageout.h>
/*
#include <mach/std_types.h> /* For pointer_t */
#include <mach/mach_types.h>
+#include <mach/mig.h>
#include <mach/kern_return.h>
#include <mach/memory_object.h>
#include <mach/memory_object_default.h>
#include <mach/memory_object_control_server.h>
-#include <mach/mach_host_server.h>
+#include <mach/host_priv_server.h>
#include <mach/boolean.h>
#include <mach/vm_prot.h>
#include <mach/message.h>
-#include <vm/vm_object.h>
-#include <vm/vm_fault.h>
/*
* Implementation dependencies:
*/
#include <string.h> /* For memcpy() */
+#include <kern/xpr.h>
+#include <kern/host.h>
+#include <kern/thread.h> /* For current_thread() */
+#include <kern/ipc_mig.h>
+#include <kern/misc_protos.h>
+
+#include <vm/vm_object.h>
+#include <vm/vm_fault.h>
#include <vm/memory_object.h>
#include <vm/vm_page.h>
#include <vm/vm_pageout.h>
#include <vm/pmap.h> /* For pmap_clear_modify */
-#include <kern/xpr.h>
-#include <kern/thread.h> /* For current_thread() */
-#include <kern/host.h>
#include <vm/vm_kern.h> /* For kernel_map, vm_move */
#include <vm/vm_map.h> /* For vm_map_pageable */
-#include <ipc/ipc_port.h>
-#include <ipc/ipc_space.h>
-
-#include <kern/misc_protos.h>
#if MACH_PAGEMAP
#include <vm/vm_external.h>
#endif /* MACH_PAGEMAP */
-ipc_port_t memory_manager_default = IP_NULL;
-vm_size_t memory_manager_default_cluster = 0;
-decl_mutex_data(,memory_manager_default_lock)
+memory_object_default_t memory_manager_default = MEMORY_OBJECT_DEFAULT_NULL;
+vm_size_t memory_manager_default_cluster = 0;
+decl_mutex_data(, memory_manager_default_lock)
/*
* Forward ref to file-local function:
*/
boolean_t
-memory_object_update(vm_object_t, vm_object_offset_t,
+vm_object_update(vm_object_t, vm_object_offset_t,
vm_size_t, memory_object_return_t, int, vm_prot_t);
*/
if (prot != VM_PROT_NO_CHANGE) {
+ if ((m->page_lock ^ prot) & prot) {
+ pmap_page_protect(m->phys_addr, VM_PROT_ALL & ~prot);
+ }
#if 0
/* code associated with the vestigial
* memory_object_data_unlock
*/
- if ((m->page_lock ^ prot) & prot) {
- pmap_page_protect(m->phys_addr, VM_PROT_ALL & ~prot);
- }
m->page_lock = prot;
m->lock_supplied = TRUE;
if (prot != VM_PROT_NONE)
return(MEMORY_OBJECT_LOCK_RESULT_DONE);
}
+
#define LIST_REQ_PAGEOUT_PAGES(object, data_cnt, action, po) \
MACRO_BEGIN \
\
\
vm_object_unlock(object); \
\
- if(((rpc_subsystem_t)pager_mux_hash_lookup(object->pager)) == \
- ((rpc_subsystem_t) &vnode_pager_workaround)) { \
- (void) vnode_pager_data_return(object->pager, \
- object->pager_request, \
- po, \
- POINTER_T(0), \
- data_cnt, \
- (action == MEMORY_OBJECT_LOCK_RESULT_MUST_CLEAN), \
- !should_flush); \
- } else { \
(void) memory_object_data_return(object->pager, \
- object->pager_request, \
po, \
- POINTER_T(0), \
data_cnt, \
(action == MEMORY_OBJECT_LOCK_RESULT_MUST_CLEAN), \
!should_flush); \
- } \
\
vm_object_lock(object); \
- \
MACRO_END
-#ifdef MACH_BSD
-#define PAGEOUT_PAGES(object, new_object, new_offset, action, po) \
-MACRO_BEGIN \
- \
- vm_map_copy_t copy; \
- register int i; \
- register vm_page_t hp; \
- \
- vm_object_unlock(object); \
- \
- (void) vm_map_copyin_object(new_object, 0, new_offset, ©); \
- \
- if(((rpc_subsystem_t)pager_mux_hash_lookup(object->pager)) == \
- ((rpc_subsystem_t) &vnode_pager_workaround)) { \
- (void) vnode_pager_data_return(object->pager, \
- object->pager_request, \
- po, \
- POINTER_T(copy), \
- new_offset, \
- (action == MEMORY_OBJECT_LOCK_RESULT_MUST_CLEAN), \
- !should_flush); \
- } else { \
- (void) memory_object_data_return(object->pager, \
- object->pager_request, \
- po, \
- POINTER_T(copy), \
- new_offset, \
- (action == MEMORY_OBJECT_LOCK_RESULT_MUST_CLEAN), \
- !should_flush); \
- } \
- \
- vm_object_lock(object); \
- \
- for (i = 0; i < atop(new_offset); i++) { \
- hp = holding_pages[i]; \
- if (hp != VM_PAGE_NULL) { \
- vm_object_paging_end(object); \
- VM_PAGE_FREE(hp); \
- } \
- } \
- \
- new_object = VM_OBJECT_NULL; \
-MACRO_END
-#else
-#define PAGEOUT_PAGES(object, new_object, new_offset, action, po) \
-MACRO_BEGIN \
- \
- vm_map_copy_t copy; \
- register int i; \
- register vm_page_t hp; \
- \
- vm_object_unlock(object); \
- \
- (void) vm_map_copyin_object(new_object, 0, new_offset, ©); \
- \
- (void) memory_object_data_return( \
- object->pager, \
- object->pager_request, \
- po, \
- POINTER_T(copy), \
- new_offset, \
- (action == MEMORY_OBJECT_LOCK_RESULT_MUST_CLEAN), \
- !should_flush); \
- \
- vm_object_lock(object); \
- \
- for (i = 0; i < atop(new_offset); i++) { \
- hp = holding_pages[i]; \
- if (hp != VM_PAGE_NULL) { \
- vm_object_paging_end(object); \
- VM_PAGE_FREE(hp); \
- } \
- } \
- \
- new_object = VM_OBJECT_NULL; \
-MACRO_END
-#endif
-
/*
* Routine: memory_object_lock_request [user interface]
*
kern_return_t
memory_object_lock_request(
- register vm_object_t object,
- register vm_object_offset_t offset,
- register vm_object_size_t size,
+ memory_object_control_t control,
+ memory_object_offset_t offset,
+ memory_object_size_t size,
memory_object_return_t should_return,
int flags,
- vm_prot_t prot,
- ipc_port_t reply_to,
- mach_msg_type_name_t reply_to_type)
+ vm_prot_t prot)
{
+ vm_object_t object;
vm_object_offset_t original_offset = offset;
boolean_t should_flush=flags & MEMORY_OBJECT_DATA_FLUSH;
XPR(XPR_MEMORY_OBJECT,
- "m_o_lock_request, obj 0x%X off 0x%X size 0x%X flags %X prot %X\n",
- (integer_t)object, offset, size,
+ "m_o_lock_request, control 0x%X off 0x%X size 0x%X flags %X prot %X\n",
+ (integer_t)control, offset, size,
(((should_return&1)<<1)|should_flush), prot);
/*
* Check for bogus arguments.
*/
+ object = memory_object_control_to_vm_object(control);
if (object == VM_OBJECT_NULL)
return (KERN_INVALID_ARGUMENT);
- if ((prot & ~VM_PROT_ALL) != 0 && prot != VM_PROT_NO_CHANGE) {
- vm_object_deallocate(object);
+ if ((prot & ~VM_PROT_ALL) != 0 && prot != VM_PROT_NO_CHANGE)
return (KERN_INVALID_ARGUMENT);
- }
size = round_page(size);
/*
* Lock the object, and acquire a paging reference to
- * prevent the memory_object and control ports from
- * being destroyed.
+ * prevent the memory_object reference from being released.
*/
-
vm_object_lock(object);
vm_object_paging_begin(object);
offset -= object->paging_offset;
- (void)memory_object_update(object,
+ (void)vm_object_update(object,
offset, size, should_return, flags, prot);
- if (IP_VALID(reply_to)) {
- vm_object_unlock(object);
-
- /* consumes our naked send-once/send right for reply_to */
- (void) memory_object_lock_completed(reply_to, reply_to_type,
- object->pager_request, original_offset, size);
-
- vm_object_lock(object);
- }
-
vm_object_paging_end(object);
vm_object_unlock(object);
- vm_object_deallocate(object);
return (KERN_SUCCESS);
}
/*
- * Routine: memory_object_sync
+ * memory_object_release_name: [interface]
+ *
+ * Enforces name semantic on memory_object reference count decrement
+ * This routine should not be called unless the caller holds a name
+ * reference gained through the memory_object_named_create or the
+ * memory_object_rename call.
+ * If the TERMINATE_IDLE flag is set, the call will return if the
+ * reference count is not 1. i.e. idle with the only remaining reference
+ * being the name.
+ * If the decision is made to proceed the name field flag is set to
+ * false and the reference count is decremented. If the RESPECT_CACHE
+ * flag is set and the reference count has gone to zero, the
+ * memory_object is checked to see if it is cacheable otherwise when
+ * the reference count is zero, it is simply terminated.
+ */
+
+kern_return_t
+memory_object_release_name(
+ memory_object_control_t control,
+ int flags)
+{
+ vm_object_t object;
+
+ object = memory_object_control_to_vm_object(control);
+ if (object == VM_OBJECT_NULL)
+ return (KERN_INVALID_ARGUMENT);
+
+ return vm_object_release_name(object, flags);
+}
+
+
+
+/*
+ * Routine: memory_object_destroy [user interface]
+ * Purpose:
+ * Shut down a memory object, despite the
+ * presence of address map (or other) references
+ * to the vm_object.
+ */
+kern_return_t
+memory_object_destroy(
+ memory_object_control_t control,
+ kern_return_t reason)
+{
+ vm_object_t object;
+
+ object = memory_object_control_to_vm_object(control);
+ if (object == VM_OBJECT_NULL)
+ return (KERN_INVALID_ARGUMENT);
+
+ return (vm_object_destroy(object, reason));
+}
+
+/*
+ * Routine: vm_object_sync
*
* Kernel internal function to synch out pages in a given
* range within an object to its memory manager. Much the
*/
boolean_t
-memory_object_sync(
+vm_object_sync(
vm_object_t object,
vm_object_offset_t offset,
- vm_object_size_t size,
+ vm_size_t size,
boolean_t should_flush,
boolean_t should_return)
{
boolean_t rv;
- XPR(XPR_MEMORY_OBJECT,
- "m_o_sync, object 0x%X, offset 0x%X size 0x%x flush %d rtn %d\n",
+ XPR(XPR_VM_OBJECT,
+ "vm_o_sync, object 0x%X, offset 0x%X size 0x%x flush %d rtn %d\n",
(integer_t)object, offset, size, should_flush, should_return);
/*
vm_object_lock(object);
vm_object_paging_begin(object);
- rv = memory_object_update(object, offset, size,
+ rv = vm_object_update(object, offset, size,
(should_return) ?
MEMORY_OBJECT_RETURN_ALL :
MEMORY_OBJECT_RETURN_NONE,
}
/*
- * Routine: memory_object_update
+ * Routine: vm_object_update
* Description:
- * Work function for m_o_lock_request(), m_o_sync().
+ * Work function for m_o_lock_request(), vm_o_sync().
*
* Called with object locked and paging ref taken.
*/
kern_return_t
-memory_object_update(
+vm_object_update(
register vm_object_t object,
register vm_object_offset_t offset,
register vm_size_t size,
boolean_t data_returned = FALSE;
boolean_t update_cow;
boolean_t should_flush = flags & MEMORY_OBJECT_DATA_FLUSH;
-#ifndef NOT_LIST_REQ
boolean_t pending_pageout = FALSE;
-#endif
/*
* To avoid blocking while scanning for pages, save
(int *)0,
&error,
FALSE,
- FALSE)) {
+ FALSE, NULL, 0)) {
case VM_FAULT_SUCCESS:
if(top_page) {
* We have completed the scan for applicable pages.
* Clean any pages that have been saved.
*/
-#ifdef NOT_LIST_REQ
- if (new_object != VM_OBJECT_NULL) {
- PAGEOUT_PAGES(object, new_object, new_offset, pageout_action,
- paging_offset);
- }
-#else
if (pending_pageout) {
LIST_REQ_PAGEOUT_PAGES(object,
data_cnt, pageout_action, paging_offset);
}
-#endif
return (data_returned);
}
kern_return_t
memory_object_synchronize_completed(
- vm_object_t object,
- vm_object_offset_t offset,
- vm_offset_t length)
+ memory_object_control_t control,
+ memory_object_offset_t offset,
+ vm_offset_t length)
{
- msync_req_t msr;
+ vm_object_t object;
+ msync_req_t msr;
XPR(XPR_MEMORY_OBJECT,
"m_o_sync_completed, object 0x%X, offset 0x%X length 0x%X\n",
* Look for bogus arguments
*/
- if (object == VM_OBJECT_NULL) {
- return KERN_INVALID_ARGUMENT;
- }
+ object = memory_object_control_to_vm_object(control);
+ if (object == VM_OBJECT_NULL)
+ return (KERN_INVALID_ARGUMENT);
vm_object_lock(object);
if (queue_end(&object->msr_q, (queue_entry_t)msr)) {
vm_object_unlock(object);
- vm_object_deallocate(object);
return KERN_INVALID_ARGUMENT;
}
msr->flag = VM_MSYNC_DONE;
msr_unlock(msr);
thread_wakeup((event_t) msr);
- vm_object_deallocate(object);
return KERN_SUCCESS;
}/* memory_object_synchronize_completed */
-
-kern_return_t
-memory_object_set_attributes_common(
+
+static kern_return_t
+vm_object_set_attributes_common(
vm_object_t object,
boolean_t may_cache,
memory_object_copy_strategy_t copy_strategy,
case MEMORY_OBJECT_COPY_DELAY:
break;
default:
- vm_object_deallocate(object);
return(KERN_INVALID_ARGUMENT);
}
#if !ADVISORY_PAGEOUT
- if (silent_overwrite || advisory_pageout) {
- vm_object_deallocate(object);
+ if (silent_overwrite || advisory_pageout)
return(KERN_INVALID_ARGUMENT);
- }
+
#endif /* !ADVISORY_PAGEOUT */
if (may_cache)
may_cache = TRUE;
* and be a power of 2 number of pages.
*/
if ((cluster_size & (PAGE_SIZE-1)) ||
- ((pages_per_cluster-1) & pages_per_cluster)) {
- vm_object_deallocate(object);
+ ((pages_per_cluster-1) & pages_per_cluster))
return KERN_INVALID_ARGUMENT;
- }
}
vm_object_lock(object);
vm_object_unlock(object);
- vm_object_deallocate(object);
-
return(KERN_SUCCESS);
}
*
* XXX This routine cannot be completed until the vm_msync, clean
* in place, and cluster work is completed. See ifdef notyet
- * below and note that memory_object_set_attributes_common()
+ * below and note that vm_object_set_attributes_common()
* may have to be expanded.
*/
kern_return_t
memory_object_change_attributes(
- vm_object_t object,
- memory_object_flavor_t flavor,
- memory_object_info_t attributes,
- mach_msg_type_number_t count,
- ipc_port_t reply_to,
- mach_msg_type_name_t reply_to_type)
+ memory_object_control_t control,
+ memory_object_flavor_t flavor,
+ memory_object_info_t attributes,
+ mach_msg_type_number_t count)
{
- kern_return_t result = KERN_SUCCESS;
- boolean_t temporary;
- boolean_t may_cache;
- boolean_t invalidate;
+ vm_object_t object;
+ kern_return_t result = KERN_SUCCESS;
+ boolean_t temporary;
+ boolean_t may_cache;
+ boolean_t invalidate;
vm_size_t cluster_size;
memory_object_copy_strategy_t copy_strategy;
- boolean_t silent_overwrite;
+ boolean_t silent_overwrite;
boolean_t advisory_pageout;
+ object = memory_object_control_to_vm_object(control);
if (object == VM_OBJECT_NULL)
- return(KERN_INVALID_ARGUMENT);
+ return (KERN_INVALID_ARGUMENT);
vm_object_lock(object);
+
temporary = object->temporary;
may_cache = object->can_persist;
copy_strategy = object->copy_strategy;
break;
}
- if (result != KERN_SUCCESS) {
- vm_object_deallocate(object);
+ if (result != KERN_SUCCESS)
return(result);
- }
if (copy_strategy == MEMORY_OBJECT_COPY_TEMPORARY) {
copy_strategy = MEMORY_OBJECT_COPY_DELAY;
}
/*
- * Do the work and throw away our object reference. It
- * is important that the object reference be deallocated
- * BEFORE sending the reply. The whole point of the reply
- * is that it shows up after the terminate message that
- * may be generated by setting the object uncacheable.
- *
* XXX may_cache may become a tri-valued variable to handle
* XXX uncache if not in use.
*/
- result = memory_object_set_attributes_common(object,
+ return (vm_object_set_attributes_common(object,
may_cache,
copy_strategy,
temporary,
cluster_size,
silent_overwrite,
- advisory_pageout);
-
- if (IP_VALID(reply_to)) {
- /* consumes our naked send-once/send right for reply_to */
- (void) memory_object_change_completed(reply_to, reply_to_type,
- object->alive ?
- object->pager_request : PAGER_REQUEST_NULL,
- flavor);
- }
-
- return(result);
+ advisory_pageout));
}
kern_return_t
memory_object_get_attributes(
- vm_object_t object,
+ memory_object_control_t control,
memory_object_flavor_t flavor,
memory_object_info_t attributes, /* pointer to OUT array */
mach_msg_type_number_t *count) /* IN/OUT */
{
- kern_return_t ret = KERN_SUCCESS;
+ kern_return_t ret = KERN_SUCCESS;
+ vm_object_t object;
- if (object == VM_OBJECT_NULL)
- return(KERN_INVALID_ARGUMENT);
+ object = memory_object_control_to_vm_object(control);
+ if (object == VM_OBJECT_NULL)
+ return (KERN_INVALID_ARGUMENT);
vm_object_lock(object);
vm_object_unlock(object);
- vm_object_deallocate(object);
-
return(ret);
}
-int vm_stat_discard_cleared_reply = 0;
-int vm_stat_discard_cleared_unset = 0;
-int vm_stat_discard_cleared_too_late = 0;
+/*
+ * Routine: memory_object_upl_request [interface]
+ * Purpose:
+ * Cause the population of a portion of a vm_object.
+ * Depending on the nature of the request, the pages
+ * returned may be contain valid data or be uninitialized.
+ *
+ */
-/*
- * vm_set_default_memory_manager():
- * [Obsolete]
+kern_return_t
+memory_object_upl_request(
+ memory_object_control_t control,
+ memory_object_offset_t offset,
+ vm_size_t size,
+ upl_t *upl_ptr,
+ upl_page_info_array_t user_page_list,
+ unsigned int *page_list_count,
+ int cntrl_flags)
+{
+ vm_object_t object;
+
+ object = memory_object_control_to_vm_object(control);
+ if (object == VM_OBJECT_NULL)
+ return (KERN_INVALID_ARGUMENT);
+
+ return vm_object_upl_request(object,
+ offset,
+ size,
+ upl_ptr,
+ user_page_list,
+ page_list_count,
+ cntrl_flags);
+}
+
+/*
+ * Routine: memory_object_super_upl_request [interface]
+ * Purpose:
+ * Cause the population of a portion of a vm_object
+ * in much the same way as memory_object_upl_request.
+ * Depending on the nature of the request, the pages
+ * returned may be contain valid data or be uninitialized.
+ * However, the region may be expanded up to the super
+ * cluster size provided.
*/
+
kern_return_t
-vm_set_default_memory_manager(
- host_t host,
- ipc_port_t *default_manager)
+memory_object_super_upl_request(
+ memory_object_control_t control,
+ memory_object_offset_t offset,
+ vm_size_t size,
+ vm_size_t super_cluster,
+ upl_t *upl,
+ upl_page_info_t *user_page_list,
+ unsigned int *page_list_count,
+ int cntrl_flags)
{
- return(host_default_memory_manager(host_priv_self(), default_manager, 4*PAGE_SIZE));
+ vm_object_t object;
+
+ object = memory_object_control_to_vm_object(control);
+ if (object == VM_OBJECT_NULL)
+ return (KERN_INVALID_ARGUMENT);
+
+ return vm_object_super_upl_request(object,
+ offset,
+ size,
+ super_cluster,
+ upl,
+ user_page_list,
+ page_list_count,
+ cntrl_flags);
}
+int vm_stat_discard_cleared_reply = 0;
+int vm_stat_discard_cleared_unset = 0;
+int vm_stat_discard_cleared_too_late = 0;
+
+
+
/*
- * Routine: host_default_memory_manager
+ * Routine: host_default_memory_manager [interface]
* Purpose:
* set/get the default memory manager port and default cluster
* size.
*/
kern_return_t
host_default_memory_manager(
- host_priv_t host_priv,
- ipc_port_t *default_manager,
- vm_size_t cluster_size)
+ host_priv_t host_priv,
+ memory_object_default_t *default_manager,
+ vm_size_t cluster_size)
{
- ipc_port_t current_manager;
- ipc_port_t new_manager;
- ipc_port_t returned_manager;
+ memory_object_default_t current_manager;
+ memory_object_default_t new_manager;
+ memory_object_default_t returned_manager;
if (host_priv == HOST_PRIV_NULL)
return(KERN_INVALID_HOST);
mutex_lock(&memory_manager_default_lock);
current_manager = memory_manager_default;
- if (new_manager == IP_NULL) {
+ if (new_manager == MEMORY_OBJECT_DEFAULT_NULL) {
/*
* Retrieve the current value.
*/
-
- returned_manager = ipc_port_copy_send(current_manager);
+ memory_object_default_reference(current_manager);
+ returned_manager = current_manager;
} else {
/*
* Retrieve the current value,
* and replace it with the supplied value.
- * We consume the supplied naked send right.
+ * We return the old reference to the caller
+ * but we have to take a reference on the new
+ * one.
*/
returned_manager = current_manager;
memory_manager_default = new_manager;
+ memory_object_default_reference(new_manager);
+
if (cluster_size % PAGE_SIZE != 0) {
#if 0
mutex_unlock(&memory_manager_default_lock);
* valid (not IP_NULL or IP_DEAD).
*/
-ipc_port_t
+__private_extern__ memory_object_default_t
memory_manager_default_reference(
vm_size_t *cluster_size)
{
- ipc_port_t current_manager;
+ memory_object_default_t current_manager;
mutex_lock(&memory_manager_default_lock);
-
- while (current_manager = ipc_port_copy_send(memory_manager_default),
- !IP_VALID(current_manager)) {
+ current_manager = memory_manager_default;
+ while (current_manager == MEMORY_OBJECT_DEFAULT_NULL) {
thread_sleep_mutex((event_t) &memory_manager_default,
&memory_manager_default_lock, THREAD_UNINT);
mutex_lock(&memory_manager_default_lock);
+ current_manager = memory_manager_default;
}
+ memory_object_default_reference(current_manager);
*cluster_size = memory_manager_default_cluster;
-
mutex_unlock(&memory_manager_default_lock);
return current_manager;
}
-/*
- * Routine: memory_manager_default_port
- * Purpose:
- * Returns true if the receiver for the port
- * is the default memory manager.
- *
- * This is a hack to let ds_read_done
- * know when it should keep memory wired.
- */
-
-boolean_t
-memory_manager_default_port(
- ipc_port_t port)
-{
- ipc_port_t current;
- boolean_t result;
-
- mutex_lock(&memory_manager_default_lock);
- current = memory_manager_default;
- if (IP_VALID(current)) {
- /*
- * There is no point in bothering to lock
- * both ports, which would be painful to do.
- * If the receive rights are moving around,
- * we might be inaccurate.
- */
-
- result = port->ip_receiver == current->ip_receiver;
- } else
- result = FALSE;
- mutex_unlock(&memory_manager_default_lock);
-
- return result;
-}
-
/*
* Routine: memory_manager_default_check
*
* but only the first time.
*
*/
-kern_return_t
+__private_extern__ kern_return_t
memory_manager_default_check(void)
{
- ipc_port_t current;
+ memory_object_default_t current;
mutex_lock(&memory_manager_default_lock);
current = memory_manager_default;
- if (!IP_VALID(current)) {
+ if (current == MEMORY_OBJECT_DEFAULT_NULL) {
static boolean_t logged; /* initialized to 0 */
boolean_t complain = !logged;
logged = TRUE;
}
}
-void
+__private_extern__ void
memory_manager_default_init(void)
{
- memory_manager_default = IP_NULL;
+ memory_manager_default = MEMORY_OBJECT_DEFAULT_NULL;
mutex_init(&memory_manager_default_lock, ETAP_VM_MEMMAN);
}
kern_return_t
memory_object_page_op(
- vm_object_t object,
- vm_object_offset_t offset,
- int ops,
- vm_offset_t *phys_entry,
- int *flags)
+ memory_object_control_t control,
+ memory_object_offset_t offset,
+ int ops,
+ vm_offset_t *phys_entry,
+ int *flags)
{
- vm_page_t dst_page;
+ vm_object_t object;
+ vm_page_t dst_page;
+
+
+ object = memory_object_control_to_vm_object(control);
+ if (object == VM_OBJECT_NULL)
+ return (KERN_INVALID_ARGUMENT);
vm_object_lock(object);
+ if(ops & UPL_POP_PHYSICAL) {
+ if(object->phys_contiguous) {
+ if (phys_entry) {
+ *phys_entry = (vm_offset_t)
+ object->shadow_offset;
+ }
+ vm_object_unlock(object);
+ return KERN_SUCCESS;
+ } else {
+ vm_object_unlock(object);
+ return KERN_INVALID_OBJECT;
+ }
+ }
+
while(TRUE) {
+ if(object->phys_contiguous) {
+ vm_object_unlock(object);
+ return KERN_INVALID_OBJECT;
+ }
+
if((dst_page = vm_page_lookup(object,offset)) == VM_PAGE_NULL) {
vm_object_unlock(object);
return KERN_FAILURE;
/* Sync up on getting the busy bit */
if((dst_page->busy || dst_page->cleaning) &&
- (((ops & UPL_POP_SET) && (ops & UPL_POP_BUSY)) || (ops & UPL_POP_DUMP))) {
+ (((ops & UPL_POP_SET) &&
+ (ops & UPL_POP_BUSY)) || (ops & UPL_POP_DUMP))) {
/* someone else is playing with the page, we will */
/* have to wait */
PAGE_ASSERT_WAIT(dst_page, THREAD_UNINT);
}
+static zone_t mem_obj_control_zone;
+
+__private_extern__ void
+memory_object_control_bootstrap(void)
+{
+ int i;
+
+ i = (vm_size_t) sizeof (struct memory_object_control);
+ mem_obj_control_zone = zinit (i, 8192*i, 4096, "mem_obj_control");
+ return;
+}
+
+__private_extern__ memory_object_control_t
+memory_object_control_allocate(
+ vm_object_t object)
+{
+ memory_object_control_t control;
+
+ control = (memory_object_control_t)zalloc(mem_obj_control_zone);
+ if (control != MEMORY_OBJECT_CONTROL_NULL)
+ control->object = object;
+ return (control);
+}
+
+__private_extern__ void
+memory_object_control_collapse(
+ memory_object_control_t control,
+ vm_object_t object)
+{
+ assert((control->object != VM_OBJECT_NULL) &&
+ (control->object != object));
+ control->object = object;
+}
+
+__private_extern__ vm_object_t
+memory_object_control_to_vm_object(
+ memory_object_control_t control)
+{
+ if (control == MEMORY_OBJECT_CONTROL_NULL)
+ return VM_OBJECT_NULL;
+
+ return (control->object);
+}
+
+memory_object_control_t
+convert_port_to_mo_control(
+ mach_port_t port)
+{
+ return MEMORY_OBJECT_CONTROL_NULL;
+}
+
+
+mach_port_t
+convert_mo_control_to_port(
+ memory_object_control_t control)
+{
+ return MACH_PORT_NULL;
+}
+
+void
+memory_object_control_reference(
+ memory_object_control_t control)
+{
+ return;
+}
+
+/*
+ * We only every issue one of these references, so kill it
+ * when that gets released (should switch the real reference
+ * counting in true port-less EMMI).
+ */
+void
+memory_object_control_deallocate(
+ memory_object_control_t control)
+{
+ zfree(mem_obj_control_zone, (vm_offset_t)control);
+}
+
+void
+memory_object_control_disable(
+ memory_object_control_t control)
+{
+ assert(control->object != VM_OBJECT_NULL);
+ control->object = VM_OBJECT_NULL;
+}
+
+void
+memory_object_default_reference(
+ memory_object_default_t dmm)
+{
+ ipc_port_make_send(dmm);
+}
+
+void
+memory_object_default_deallocate(
+ memory_object_default_t dmm)
+{
+ ipc_port_release_send(dmm);
+}
+
+memory_object_t
+convert_port_to_memory_object(
+ mach_port_t port)
+{
+ return (MEMORY_OBJECT_NULL);
+}
+
+
+mach_port_t
+convert_memory_object_to_port(
+ memory_object_t object)
+{
+ return (MACH_PORT_NULL);
+}
+
+#ifdef MACH_BSD
+/* remove after component interface available */
+extern int vnode_pager_workaround;
+extern int device_pager_workaround;
+#endif
+
+
+/* Routine memory_object_reference */
+void memory_object_reference(
+ memory_object_t memory_object)
+{
+extern void dp_memory_object_reference(memory_object_t);
+
+#ifdef MACH_BSD
+ extern void vnode_pager_reference(memory_object_t);
+ extern void device_pager_reference(memory_object_t);
+
+ if(memory_object->pager == &vnode_pager_workaround) {
+ vnode_pager_reference(memory_object);
+ } else if(memory_object->pager == &device_pager_workaround) {
+ device_pager_reference(memory_object);
+ } else
+#endif
+ dp_memory_object_reference(memory_object);
+}
+
+/* Routine memory_object_deallocate */
+void memory_object_deallocate(
+ memory_object_t memory_object)
+{
+extern void dp_memory_object_deallocate(memory_object_t);
+
+#ifdef MACH_BSD
+ extern void vnode_pager_deallocate(memory_object_t);
+ extern void device_pager_deallocate(memory_object_t);
+
+ if(memory_object->pager == &vnode_pager_workaround) {
+ vnode_pager_deallocate(memory_object);
+ } else if(memory_object->pager == &device_pager_workaround) {
+ device_pager_deallocate(memory_object);
+ } else
+#endif
+ dp_memory_object_deallocate(memory_object);
+}
+
+
+/* Routine memory_object_init */
+kern_return_t memory_object_init
+(
+ memory_object_t memory_object,
+ memory_object_control_t memory_control,
+ vm_size_t memory_object_page_size
+)
+{
+extern kern_return_t dp_memory_object_init(memory_object_t,
+ memory_object_control_t,
+ vm_size_t);
+#ifdef MACH_BSD
+extern kern_return_t vnode_pager_init(memory_object_t,
+ memory_object_control_t,
+ vm_size_t);
+extern kern_return_t device_pager_init(memory_object_t,
+ memory_object_control_t,
+ vm_size_t);
+
+ if(memory_object->pager == &vnode_pager_workaround) {
+ return vnode_pager_init(memory_object,
+ memory_control,
+ memory_object_page_size);
+ } else if(memory_object->pager == &device_pager_workaround) {
+ return device_pager_init(memory_object,
+ memory_control,
+ memory_object_page_size);
+ } else
+#endif
+ return dp_memory_object_init(memory_object,
+ memory_control,
+ memory_object_page_size);
+}
+
+/* Routine memory_object_terminate */
+kern_return_t memory_object_terminate
+(
+ memory_object_t memory_object
+)
+{
+extern kern_return_t dp_memory_object_terminate(memory_object_t);
+
+#ifdef MACH_BSD
+extern kern_return_t vnode_pager_terminate(memory_object_t);
+extern kern_return_t device_pager_terminate(memory_object_t);
+
+ if(memory_object->pager == &vnode_pager_workaround) {
+ return vnode_pager_terminate(memory_object);
+ } else if(memory_object->pager == &device_pager_workaround) {
+ return device_pager_terminate(memory_object);
+ } else
+#endif
+ return dp_memory_object_terminate(memory_object);
+}
+
+/* Routine memory_object_data_request */
+kern_return_t memory_object_data_request
+(
+ memory_object_t memory_object,
+ memory_object_offset_t offset,
+ vm_size_t length,
+ vm_prot_t desired_access
+)
+{
+extern kern_return_t dp_memory_object_data_request(memory_object_t,
+ memory_object_offset_t, vm_size_t, vm_prot_t);
+
+#ifdef MACH_BSD
+extern kern_return_t vnode_pager_data_request(memory_object_t,
+ memory_object_offset_t, vm_size_t, vm_prot_t);
+extern kern_return_t device_pager_data_request(memory_object_t,
+ memory_object_offset_t, vm_size_t, vm_prot_t);
+
+ if (memory_object->pager == &vnode_pager_workaround) {
+ return vnode_pager_data_request(memory_object,
+ offset,
+ length,
+ desired_access);
+ } else if (memory_object->pager == &device_pager_workaround) {
+ return device_pager_data_request(memory_object,
+ offset,
+ length,
+ desired_access);
+ } else
+#endif
+ return dp_memory_object_data_request(memory_object,
+ offset,
+ length,
+ desired_access);
+}
+
+/* Routine memory_object_data_return */
+kern_return_t memory_object_data_return
+(
+ memory_object_t memory_object,
+ memory_object_offset_t offset,
+ vm_size_t size,
+ boolean_t dirty,
+ boolean_t kernel_copy
+)
+{
+ extern kern_return_t dp_memory_object_data_return(memory_object_t,
+ memory_object_offset_t,
+ vm_size_t,
+ boolean_t,
+ boolean_t);
+#ifdef MACH_BSD
+ extern kern_return_t vnode_pager_data_return(memory_object_t,
+ memory_object_offset_t,
+ vm_size_t,
+ boolean_t,
+ boolean_t);
+ extern kern_return_t device_pager_data_return(memory_object_t,
+ memory_object_offset_t,
+ vm_size_t,
+ boolean_t,
+ boolean_t);
+
+ if (memory_object->pager == &vnode_pager_workaround) {
+ return vnode_pager_data_return(memory_object,
+ offset,
+ size,
+ dirty,
+ kernel_copy);
+ } else if (memory_object->pager == &device_pager_workaround) {
+ return device_pager_data_return(memory_object,
+ offset,
+ size,
+ dirty,
+ kernel_copy);
+ } else
+#endif
+ return dp_memory_object_data_return(memory_object,
+ offset,
+ size,
+ dirty,
+ kernel_copy);
+}
+
+/* Routine memory_object_data_initialize */
+kern_return_t memory_object_data_initialize
+(
+ memory_object_t memory_object,
+ memory_object_offset_t offset,
+ vm_size_t size
+)
+{
+
+ extern kern_return_t dp_memory_object_data_initialize(memory_object_t,
+ memory_object_offset_t,
+ vm_size_t);
+#ifdef MACH_BSD
+ extern kern_return_t vnode_pager_data_initialize(memory_object_t,
+ memory_object_offset_t,
+ vm_size_t);
+ extern kern_return_t device_pager_data_initialize(memory_object_t,
+ memory_object_offset_t,
+ vm_size_t);
+
+ if (memory_object->pager == &vnode_pager_workaround) {
+ return vnode_pager_data_initialize(memory_object,
+ offset,
+ size);
+ } else if (memory_object->pager == &device_pager_workaround) {
+ return device_pager_data_initialize(memory_object,
+ offset,
+ size);
+ } else
+#endif
+ return dp_memory_object_data_initialize(memory_object,
+ offset,
+ size);
+}
+
+/* Routine memory_object_data_unlock */
+kern_return_t memory_object_data_unlock
+(
+ memory_object_t memory_object,
+ memory_object_offset_t offset,
+ vm_size_t size,
+ vm_prot_t desired_access
+)
+{
+ extern kern_return_t dp_memory_object_data_unlock(memory_object_t,
+ memory_object_offset_t,
+ vm_size_t,
+ vm_prot_t);
+#ifdef MACH_BSD
+ extern kern_return_t vnode_pager_data_unlock(memory_object_t,
+ memory_object_offset_t,
+ vm_size_t,
+ vm_prot_t);
+ extern kern_return_t device_pager_data_unlock(memory_object_t,
+ memory_object_offset_t,
+ vm_size_t,
+ vm_prot_t);
+
+ if (memory_object->pager == &vnode_pager_workaround) {
+ return vnode_pager_data_unlock(memory_object,
+ offset,
+ size,
+ desired_access);
+ } else if (memory_object->pager == &device_pager_workaround) {
+ return device_pager_data_unlock(memory_object,
+ offset,
+ size,
+ desired_access);
+ } else
+#endif
+ return dp_memory_object_data_unlock(memory_object,
+ offset,
+ size,
+ desired_access);
+
+}
+
+/* Routine memory_object_synchronize */
+kern_return_t memory_object_synchronize
+(
+ memory_object_t memory_object,
+ memory_object_offset_t offset,
+ vm_size_t size,
+ vm_sync_t sync_flags
+)
+{
+ extern kern_return_t dp_memory_object_data_synchronize(memory_object_t,
+ memory_object_offset_t,
+ vm_size_t,
+ vm_sync_t);
+#ifdef MACH_BSD
+ extern kern_return_t vnode_pager_data_synchronize(memory_object_t,
+ memory_object_offset_t,
+ vm_size_t,
+ vm_sync_t);
+ extern kern_return_t device_pager_data_synchronize(memory_object_t,
+ memory_object_offset_t,
+ vm_size_t,
+ vm_sync_t);
+
+ if (memory_object->pager == &vnode_pager_workaround) {
+ return vnode_pager_synchronize(
+ memory_object,
+ offset,
+ size,
+ sync_flags);
+ } else if (memory_object->pager == &device_pager_workaround) {
+ return device_pager_synchronize(
+ memory_object,
+ offset,
+ size,
+ sync_flags);
+ } else
+#endif
+ return dp_memory_object_synchronize(
+ memory_object,
+ offset,
+ size,
+ sync_flags);
+}
+
+/* Routine memory_object_unmap */
+kern_return_t memory_object_unmap
+(
+ memory_object_t memory_object
+)
+{
+ extern kern_return_t dp_memory_object_unmap(memory_object_t);
+#ifdef MACH_BSD
+ extern kern_return_t vnode_pager_unmap(memory_object_t);
+ extern kern_return_t device_pager_unmap(memory_object_t);
+
+ if (memory_object->pager == &vnode_pager_workaround) {
+ return vnode_pager_unmap(memory_object);
+ } else if (memory_object->pager == &device_pager_workaround) {
+ return device_pager_unmap(memory_object);
+ } else
+#endif
+ return dp_memory_object_unmap(memory_object);
+}
+
+/* Routine memory_object_create */
+kern_return_t memory_object_create
+(
+ memory_object_default_t default_memory_manager,
+ vm_size_t new_memory_object_size,
+ memory_object_t *new_memory_object
+)
+{
+extern kern_return_t default_pager_memory_object_create(memory_object_default_t,
+ vm_size_t,
+ memory_object_t *);
+
+ return default_pager_memory_object_create(default_memory_manager,
+ new_memory_object_size,
+ new_memory_object);
+}