/*
- * Copyright (c) 1998-2004 Apple Computer, Inc. All rights reserved.
+ * Copyright (c) 1998-2014 Apple Inc. All rights reserved.
*
* @APPLE_OSREFERENCE_LICENSE_HEADER_START@
*
*/
+#include <libkern/c++/OSKext.h>
#include <IOKit/IOKitServer.h>
#include <IOKit/IOKitKeysPrivate.h>
#include <IOKit/IOUserClient.h>
#include <IOKit/IORegistryEntry.h>
#include <IOKit/IOCatalogue.h>
#include <IOKit/IOMemoryDescriptor.h>
+#include <IOKit/IOBufferMemoryDescriptor.h>
#include <IOKit/IOLib.h>
+#include <IOKit/IOStatisticsPrivate.h>
+#include <IOKit/IOTimeStamp.h>
+#include <IOKit/system.h>
+#include <libkern/OSDebug.h>
+#include <sys/proc.h>
+#include <sys/kauth.h>
+#include <sys/codesign.h>
+
+#if CONFIG_MACF
+
+extern "C" {
+#include <security/mac_framework.h>
+};
+#include <sys/kauth.h>
+
+#define IOMACF_LOG 0
+
+#endif /* CONFIG_MACF */
#include <IOKit/assert.h>
#include "IOServicePrivate.h"
+#include "IOKitKernelInternal.h"
+
+#define SCALAR64(x) ((io_user_scalar_t)((unsigned int)x))
+#define SCALAR32(x) ((uint32_t )x)
+#define ARG32(x) ((void *)(uintptr_t)SCALAR32(x))
+#define REF64(x) ((io_user_reference_t)((UInt64)(x)))
+#define REF32(x) ((int)(x))
+
+enum
+{
+ kIOUCAsync0Flags = 3ULL,
+ kIOUCAsync64Flag = 1ULL
+};
+
+#if IOKITSTATS
+
+#define IOStatisticsRegisterCounter() \
+do { \
+ reserved->counter = IOStatistics::registerUserClient(this); \
+} while (0)
+
+#define IOStatisticsUnregisterCounter() \
+do { \
+ if (reserved) \
+ IOStatistics::unregisterUserClient(reserved->counter); \
+} while (0)
+
+#define IOStatisticsClientCall() \
+do { \
+ IOStatistics::countUserClientCall(client); \
+} while (0)
+
+#else
+
+#define IOStatisticsRegisterCounter()
+#define IOStatisticsUnregisterCounter()
+#define IOStatisticsClientCall()
+
+#endif /* IOKITSTATS */
/* * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * */
extern void iokit_retain_port( ipc_port_t port );
extern void iokit_release_port( ipc_port_t port );
+extern void iokit_release_port_send( ipc_port_t port );
extern kern_return_t iokit_switch_object_port( ipc_port_t port, io_object_t obj, ipc_kobject_type_t type );
machPort = (IOMachPort *) dict->getObject( (const OSSymbol *) obj );
if( machPort) {
- destroyed = (machPort->mscount == *mscount);
+ destroyed = (machPort->mscount <= *mscount);
if( destroyed)
dict->removeObject( (const OSSymbol *) obj );
else
};
/* * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * */
-
extern "C" {
// functions called from osfmk/device/iokit_rpc.c
if( IKOT_IOKIT_CONNECT == type)
{
- if( (client = OSDynamicCast( IOUserClient, obj )))
+ if( (client = OSDynamicCast( IOUserClient, obj ))) {
+ IOStatisticsClientCall();
client->clientDied();
}
+ }
else if( IKOT_IOKIT_OBJECT == type)
{
if( (map = OSDynamicCast( IOMemoryMap, obj )))
struct PingMsg {
mach_msg_header_t msgHdr;
- OSNotificationHeader notifyHeader;
+ OSNotificationHeader64 notifyHeader;
};
enum { kMaxOutstanding = 1024 };
- PingMsg * pingMsg;
+ PingMsg * pingMsg;
vm_size_t msgSize;
OSArray * newSet;
OSObject * lastEntry;
public:
virtual bool init( mach_port_t port, natural_t type,
- OSAsyncReference reference );
+ void * reference, vm_size_t referenceSize,
+ bool clientIs64 );
virtual void free();
static bool _handler( void * target,
- void * ref, IOService * newService );
+ void * ref, IOService * newService, IONotifier * notifier );
virtual bool handler( void * ref, IOService * newService );
virtual OSObject * getNextObject();
mach_msg_header_t msgHdr;
mach_msg_body_t msgBody;
mach_msg_port_descriptor_t ports[1];
- OSNotificationHeader notifyHeader;
+ OSNotificationHeader64 notifyHeader __attribute__ ((packed));
};
PingMsg * pingMsg;
vm_size_t msgSize;
+ uint8_t clientIs64;
+ int owningPID;
public:
virtual bool init( mach_port_t port, natural_t type,
- OSAsyncReference reference, vm_size_t extraSize );
+ void * reference, vm_size_t referenceSize,
+ vm_size_t extraSize,
+ bool clientIs64 );
+
virtual void free();
static IOReturn _handler( void * target, void * ref,
/* * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * */
bool IOServiceUserNotification::init( mach_port_t port, natural_t type,
- OSAsyncReference reference )
+ void * reference, vm_size_t referenceSize,
+ bool clientIs64 )
{
+ if( !super::init())
+ return( false );
+
newSet = OSArray::withCapacity( 1 );
if( !newSet)
return( false );
- msgSize = sizeof( PingMsg) + 0;
+ if (referenceSize > sizeof(OSAsyncReference64))
+ return( false );
+
+ msgSize = sizeof(PingMsg) - sizeof(OSAsyncReference64) + referenceSize;
pingMsg = (PingMsg *) IOMalloc( msgSize);
if( !pingMsg)
return( false );
pingMsg->notifyHeader.size = 0;
pingMsg->notifyHeader.type = type;
- bcopy( reference, pingMsg->notifyHeader.reference, sizeof(OSAsyncReference) );
+ bcopy( reference, pingMsg->notifyHeader.reference, referenceSize );
- return( super::init() );
+ return( true );
}
void IOServiceUserNotification::free( void )
super::free();
- if( _pingMsg && _msgSize)
- IOFree( _pingMsg, _msgSize);
+ if( _pingMsg && _msgSize) {
+ if (_pingMsg->msgHdr.msgh_remote_port) {
+ iokit_release_port_send(_pingMsg->msgHdr.msgh_remote_port);
+ }
+ IOFree(_pingMsg, _msgSize);
+ }
if( _lastEntry)
_lastEntry->release();
}
bool IOServiceUserNotification::_handler( void * target,
- void * ref, IOService * newService )
+ void * ref, IOService * newService, IONotifier * notifier )
{
return( ((IOServiceUserNotification *) target)->handler( ref, newService ));
}
else
pingMsg->msgHdr.msgh_local_port = NULL;
- kr = mach_msg_send_from_kernel( &pingMsg->msgHdr,
- pingMsg->msgHdr.msgh_size);
+ kr = mach_msg_send_from_kernel_with_options( &pingMsg->msgHdr,
+ pingMsg->msgHdr.msgh_size,
+ (MACH_SEND_MSG | MACH_SEND_ALWAYS | MACH_SEND_IMPORTANCE),
+ 0);
if( port)
iokit_release_port( port );
if( KERN_SUCCESS != kr)
- IOLog("%s: mach_msg_send_from_kernel {%x}\n", __FILE__, kr );
+ IOLog("%s: mach_msg_send_from_kernel_proper {%x}\n", __FILE__, kr );
}
return( true );
/* * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * */
bool IOServiceMessageUserNotification::init( mach_port_t port, natural_t type,
- OSAsyncReference reference, vm_size_t extraSize )
+ void * reference, vm_size_t referenceSize, vm_size_t extraSize,
+ bool client64 )
{
+ if( !super::init())
+ return( false );
- extraSize += sizeof(IOServiceInterestContent);
- msgSize = sizeof( PingMsg) + extraSize;
+ if (referenceSize > sizeof(OSAsyncReference64))
+ return( false );
+
+ clientIs64 = client64;
+
+ owningPID = proc_selfpid();
+
+ extraSize += sizeof(IOServiceInterestContent64);
+ msgSize = sizeof(PingMsg) - sizeof(OSAsyncReference64) + referenceSize + extraSize;
pingMsg = (PingMsg *) IOMalloc( msgSize);
if( !pingMsg)
return( false );
pingMsg->notifyHeader.size = extraSize;
pingMsg->notifyHeader.type = type;
- bcopy( reference, pingMsg->notifyHeader.reference, sizeof(OSAsyncReference) );
+ bcopy( reference, pingMsg->notifyHeader.reference, referenceSize );
- return( super::init() );
+ return( true );
}
void IOServiceMessageUserNotification::free( void )
super::free();
- if( _pingMsg && _msgSize)
+ if( _pingMsg && _msgSize) {
+ if (_pingMsg->msgHdr.msgh_remote_port) {
+ iokit_release_port_send(_pingMsg->msgHdr.msgh_remote_port);
+ }
IOFree( _pingMsg, _msgSize);
+ }
}
IOReturn IOServiceMessageUserNotification::_handler( void * target, void * ref,
UInt32 messageType, IOService * provider,
void * messageArgument, vm_size_t argSize )
{
- kern_return_t kr;
- ipc_port_t thisPort, providerPort;
- IOServiceInterestContent * data = (IOServiceInterestContent *)
- pingMsg->notifyHeader.content;
+ kern_return_t kr;
+ ipc_port_t thisPort, providerPort;
+ IOServiceInterestContent64 * data = (IOServiceInterestContent64 *)
+ ((((uint8_t *) pingMsg) + msgSize) - pingMsg->notifyHeader.size);
+ // == pingMsg->notifyHeader.content;
+
+ if (kIOMessageCopyClientID == messageType)
+ {
+ *((void **) messageArgument) = OSNumber::withNumber(owningPID, 32);
+ return (kIOReturnSuccess);
+ }
data->messageType = messageType;
- if( argSize == 0) {
- argSize = sizeof( messageArgument);
- data->messageArgument[0] = messageArgument;
- } else {
+
+ if( argSize == 0)
+ {
+ data->messageArgument[0] = (io_user_reference_t) messageArgument;
+ if (clientIs64)
+ argSize = sizeof(data->messageArgument[0]);
+ else
+ {
+ data->messageArgument[0] |= (data->messageArgument[0] << 32);
+ argSize = sizeof(uint32_t);
+ }
+ }
+ else
+ {
if( argSize > kIOUserNotifyMaxMessageSize)
argSize = kIOUserNotifyMaxMessageSize;
bcopy( messageArgument, data->messageArgument, argSize );
}
- pingMsg->msgHdr.msgh_size = sizeof( PingMsg)
- + sizeof( IOServiceInterestContent )
+
+ // adjust message size for ipc restrictions
+ natural_t type;
+ type = pingMsg->notifyHeader.type;
+ type &= ~(kIOKitNoticationMsgSizeMask << kIOKitNoticationTypeSizeAdjShift);
+ type |= ((argSize & kIOKitNoticationMsgSizeMask) << kIOKitNoticationTypeSizeAdjShift);
+ pingMsg->notifyHeader.type = type;
+ argSize = (argSize + kIOKitNoticationMsgSizeMask) & ~kIOKitNoticationMsgSizeMask;
+
+ pingMsg->msgHdr.msgh_size = msgSize - pingMsg->notifyHeader.size
+ + sizeof( IOServiceInterestContent64 )
- sizeof( data->messageArgument)
+ argSize;
pingMsg->ports[0].name = providerPort;
thisPort = iokit_port_for_object( this, IKOT_IOKIT_OBJECT );
pingMsg->msgHdr.msgh_local_port = thisPort;
- kr = mach_msg_send_from_kernel( &pingMsg->msgHdr,
- pingMsg->msgHdr.msgh_size);
+ kr = mach_msg_send_from_kernel_with_options( &pingMsg->msgHdr,
+ pingMsg->msgHdr.msgh_size,
+ (MACH_SEND_MSG | MACH_SEND_ALWAYS | MACH_SEND_IMPORTANCE),
+ 0);
if( thisPort)
iokit_release_port( thisPort );
if( providerPort)
iokit_release_port( providerPort );
if( KERN_SUCCESS != kr)
- IOLog("%s: mach_msg_send_from_kernel {%x}\n", __FILE__, kr );
+ IOLog("%s: mach_msg_send_from_kernel_proper {%x}\n", __FILE__, kr );
return( kIOReturnSuccess );
}
mach_port_t wakePort,
void *callback, void *refcon)
{
- asyncRef[kIOAsyncReservedIndex] = (natural_t) wakePort;
- asyncRef[kIOAsyncCalloutFuncIndex] = (natural_t) callback;
- asyncRef[kIOAsyncCalloutRefconIndex] = (natural_t) refcon;
+ asyncRef[kIOAsyncReservedIndex] = ((uintptr_t) wakePort)
+ | (kIOUCAsync0Flags & asyncRef[kIOAsyncReservedIndex]);
+ asyncRef[kIOAsyncCalloutFuncIndex] = (uintptr_t) callback;
+ asyncRef[kIOAsyncCalloutRefconIndex] = (uintptr_t) refcon;
+}
+
+void IOUserClient::setAsyncReference64(OSAsyncReference64 asyncRef,
+ mach_port_t wakePort,
+ mach_vm_address_t callback, io_user_reference_t refcon)
+{
+ asyncRef[kIOAsyncReservedIndex] = ((io_user_reference_t) wakePort)
+ | (kIOUCAsync0Flags & asyncRef[kIOAsyncReservedIndex]);
+ asyncRef[kIOAsyncCalloutFuncIndex] = (io_user_reference_t) callback;
+ asyncRef[kIOAsyncCalloutRefconIndex] = refcon;
+}
+
+void IOUserClient::setAsyncReference64(OSAsyncReference64 asyncRef,
+ mach_port_t wakePort,
+ mach_vm_address_t callback, io_user_reference_t refcon, task_t task)
+{
+ setAsyncReference64(asyncRef, wakePort, callback, refcon);
+ if (vm_map_is_64bit(get_task_map(task))) {
+ asyncRef[kIOAsyncReservedIndex] |= kIOUCAsync64Flag;
+ }
}
-inline OSDictionary * CopyConsoleUser(UInt32 uid)
+static OSDictionary * CopyConsoleUser(UInt32 uid)
{
OSArray * array;
OSDictionary * user = 0;
return user;
}
+static OSDictionary * CopyUserOnConsole(void)
+{
+ OSArray * array;
+ OSDictionary * user = 0;
+
+ if ((array = OSDynamicCast(OSArray,
+ IORegistryEntry::getRegistryRoot()->copyProperty(gIOConsoleUsersKey))))
+ {
+ for (unsigned int idx = 0;
+ (user = OSDynamicCast(OSDictionary, array->getObject(idx)));
+ idx++)
+ {
+ if (kOSBooleanTrue == user->getObject(gIOConsoleSessionOnConsoleKey))
+ {
+ user->retain();
+ break;
+ }
+ }
+ array->release();
+ }
+ return (user);
+}
+
+IOReturn IOUserClient::clientHasAuthorization( task_t task,
+ IOService * service )
+{
+ proc_t p;
+
+ p = (proc_t) get_bsdtask_info(task);
+ if (p)
+ {
+ uint64_t authorizationID;
+
+ authorizationID = proc_uniqueid(p);
+ if (authorizationID)
+ {
+ if (service->getAuthorizationID() == authorizationID)
+ {
+ return (kIOReturnSuccess);
+ }
+ }
+ }
+
+ return (kIOReturnNotPermitted);
+}
+
IOReturn IOUserClient::clientHasPrivilege( void * securityToken,
const char * privilegeName )
{
OSDictionary * user;
bool secureConsole;
- if ((secureConsole = !strcmp(privilegeName, kIOClientPrivilegeSecureConsoleProcess)))
+
+ if (!strncmp(privilegeName, kIOClientPrivilegeForeground,
+ sizeof(kIOClientPrivilegeForeground)))
+ {
+ if (task_is_gpu_denied(current_task()))
+ return (kIOReturnNotPrivileged);
+ else
+ return (kIOReturnSuccess);
+ }
+
+ if (!strncmp(privilegeName, kIOClientPrivilegeConsoleSession,
+ sizeof(kIOClientPrivilegeConsoleSession)))
+ {
+ kauth_cred_t cred;
+ proc_t p;
+
+ task = (task_t) securityToken;
+ if (!task)
+ task = current_task();
+ p = (proc_t) get_bsdtask_info(task);
+ kr = kIOReturnNotPrivileged;
+
+ if (p && (cred = kauth_cred_proc_ref(p)))
+ {
+ user = CopyUserOnConsole();
+ if (user)
+ {
+ OSNumber * num;
+ if ((num = OSDynamicCast(OSNumber, user->getObject(gIOConsoleSessionAuditIDKey)))
+ && (cred->cr_audit.as_aia_p->ai_asid == (au_asid_t) num->unsigned32BitValue()))
+ {
+ kr = kIOReturnSuccess;
+ }
+ user->release();
+ }
+ kauth_cred_unref(&cred);
+ }
+ return (kr);
+ }
+
+ if ((secureConsole = !strncmp(privilegeName, kIOClientPrivilegeSecureConsoleProcess,
+ sizeof(kIOClientPrivilegeSecureConsoleProcess))))
task = (task_t)((IOUCProcessToken *)securityToken)->token;
else
task = (task_t)securityToken;
-
+
count = TASK_SECURITY_TOKEN_COUNT;
kr = task_info( task, TASK_SECURITY_TOKEN, (task_info_t) &token, &count );
if (KERN_SUCCESS != kr)
{}
- else if (!strcmp(privilegeName, kIOClientPrivilegeAdministrator)) {
+ else if (!strncmp(privilegeName, kIOClientPrivilegeAdministrator,
+ sizeof(kIOClientPrivilegeAdministrator))) {
if (0 != token.val[0])
kr = kIOReturnNotPrivileged;
- } else if (!strcmp(privilegeName, kIOClientPrivilegeLocalUser)) {
+ } else if (!strncmp(privilegeName, kIOClientPrivilegeLocalUser,
+ sizeof(kIOClientPrivilegeLocalUser))) {
user = CopyConsoleUser(token.val[0]);
if ( user )
user->release();
else
kr = kIOReturnNotPrivileged;
- } else if (secureConsole || !strcmp(privilegeName, kIOClientPrivilegeConsoleUser)) {
+ } else if (secureConsole || !strncmp(privilegeName, kIOClientPrivilegeConsoleUser,
+ sizeof(kIOClientPrivilegeConsoleUser))) {
user = CopyConsoleUser(token.val[0]);
if ( user ) {
if (user->getObject(gIOConsoleSessionOnConsoleKey) != kOSBooleanTrue)
return (kr);
}
+OSObject * IOUserClient::copyClientEntitlement( task_t task,
+ const char * entitlement )
+{
+#define MAX_ENTITLEMENTS_LEN (128 * 1024)
+
+ proc_t p = NULL;
+ pid_t pid = 0;
+ char procname[MAXCOMLEN + 1] = "";
+ size_t len = 0;
+ void *entitlements_blob = NULL;
+ char *entitlements_data = NULL;
+ OSObject *entitlements_obj = NULL;
+ OSDictionary *entitlements = NULL;
+ OSString *errorString = NULL;
+ OSObject *value = NULL;
+
+ p = (proc_t)get_bsdtask_info(task);
+ if (p == NULL)
+ goto fail;
+ pid = proc_pid(p);
+ proc_name(pid, procname, (int)sizeof(procname));
+
+ if (cs_entitlements_blob_get(p, &entitlements_blob, &len) != 0)
+ goto fail;
+
+ if (len <= offsetof(CS_GenericBlob, data))
+ goto fail;
+
+ /*
+ * Per <rdar://problem/11593877>, enforce a limit on the amount of XML
+ * we'll try to parse in the kernel.
+ */
+ len -= offsetof(CS_GenericBlob, data);
+ if (len > MAX_ENTITLEMENTS_LEN) {
+ IOLog("failed to parse entitlements for %s[%u]: %lu bytes of entitlements exceeds maximum of %u\n", procname, pid, len, MAX_ENTITLEMENTS_LEN);
+ goto fail;
+ }
+
+ /*
+ * OSUnserializeXML() expects a nul-terminated string, but that isn't
+ * what is stored in the entitlements blob. Copy the string and
+ * terminate it.
+ */
+ entitlements_data = (char *)IOMalloc(len + 1);
+ if (entitlements_data == NULL)
+ goto fail;
+ memcpy(entitlements_data, ((CS_GenericBlob *)entitlements_blob)->data, len);
+ entitlements_data[len] = '\0';
+
+ entitlements_obj = OSUnserializeXML(entitlements_data, len + 1, &errorString);
+ if (errorString != NULL) {
+ IOLog("failed to parse entitlements for %s[%u]: %s\n", procname, pid, errorString->getCStringNoCopy());
+ goto fail;
+ }
+ if (entitlements_obj == NULL)
+ goto fail;
+
+ entitlements = OSDynamicCast(OSDictionary, entitlements_obj);
+ if (entitlements == NULL)
+ goto fail;
+
+ /* Fetch the entitlement value from the dictionary. */
+ value = entitlements->getObject(entitlement);
+ if (value != NULL)
+ value->retain();
+
+fail:
+ if (entitlements_data != NULL)
+ IOFree(entitlements_data, len + 1);
+ if (entitlements_obj != NULL)
+ entitlements_obj->release();
+ if (errorString != NULL)
+ errorString->release();
+ return value;
+}
+
bool IOUserClient::init()
{
- if( getPropertyTable())
- return true;
- else
- return super::init();
+ if (getPropertyTable() || super::init())
+ return reserve();
+
+ return false;
}
bool IOUserClient::init(OSDictionary * dictionary)
{
- if( getPropertyTable())
- return true;
- else
- return super::init(dictionary);
+ if (getPropertyTable() || super::init(dictionary))
+ return reserve();
+
+ return false;
}
bool IOUserClient::initWithTask(task_t owningTask,
void * securityID,
UInt32 type )
-{
- if( getPropertyTable())
- return true;
- else
- return super::init();
+{
+ if (getPropertyTable() || super::init())
+ return reserve();
+
+ return false;
}
bool IOUserClient::initWithTask(task_t owningTask,
return( ok );
}
+bool IOUserClient::reserve()
+{
+ if(!reserved) {
+ reserved = IONew(ExpansionData, 1);
+ if (!reserved) {
+ return false;
+ }
+ }
+ setTerminateDefer(NULL, true);
+ IOStatisticsRegisterCounter();
+
+ return true;
+}
+
void IOUserClient::free()
{
if( mappings)
mappings->release();
+
+ IOStatisticsUnregisterCounter();
+ if (reserved)
+ IODelete(reserved, ExpansionData, 1);
+
super::free();
}
return( kIOReturnUnsupported);
}
+IOReturn IOUserClient::registerNotificationPort(
+ mach_port_t port,
+ UInt32 type,
+ io_user_reference_t refCon)
+{
+ return (registerNotificationPort(port, type, (UInt32) refCon));
+}
+
IOReturn IOUserClient::getNotificationSemaphore( UInt32 notification_type,
semaphore_t * semaphore )
{
return( kIOReturnUnsupported);
}
+#if !__LP64__
IOMemoryMap * IOUserClient::mapClientMemory(
IOOptionBits type,
task_t task,
IOOptionBits mapFlags,
IOVirtualAddress atAddress )
+{
+ return (NULL);
+}
+#endif
+
+IOMemoryMap * IOUserClient::mapClientMemory64(
+ IOOptionBits type,
+ task_t task,
+ IOOptionBits mapFlags,
+ mach_vm_address_t atAddress )
{
IOReturn err;
IOOptionBits options = 0;
options = (options & ~kIOMapUserOptionsMask)
| (mapFlags & kIOMapUserOptionsMask);
- map = memory->map( task, atAddress, options );
+ map = memory->createMappingInTask( task, atAddress, options );
memory->release();
}
mach_port_name_t name;
name = IOMachPort::makeSendRightForTask( task, obj, IKOT_IOKIT_OBJECT );
- assert( name );
*(mach_port_name_t *)clientObj = name;
return kIOReturnSuccess;
return trap;
}
+IOReturn IOUserClient::releaseAsyncReference64(OSAsyncReference64 reference)
+{
+ mach_port_t port;
+ port = (mach_port_t) (reference[0] & ~kIOUCAsync0Flags);
+
+ if (MACH_PORT_NULL != port)
+ iokit_release_port_send(port);
+
+ return (kIOReturnSuccess);
+}
+
+IOReturn IOUserClient::releaseNotificationPort(mach_port_t port)
+{
+ if (MACH_PORT_NULL != port)
+ iokit_release_port_send(port);
+
+ return (kIOReturnSuccess);
+}
+
IOReturn IOUserClient::sendAsyncResult(OSAsyncReference reference,
IOReturn result, void *args[], UInt32 numArgs)
{
- struct ReplyMsg {
- mach_msg_header_t msgHdr;
- OSNotificationHeader notifyHdr;
- IOAsyncCompletionContent asyncContent;
- void * args[kMaxAsyncArgs];
+ OSAsyncReference64 reference64;
+ io_user_reference_t args64[kMaxAsyncArgs];
+ unsigned int idx;
+
+ if (numArgs > kMaxAsyncArgs)
+ return kIOReturnMessageTooLarge;
+
+ for (idx = 0; idx < kOSAsyncRef64Count; idx++)
+ reference64[idx] = REF64(reference[idx]);
+
+ for (idx = 0; idx < numArgs; idx++)
+ args64[idx] = REF64(args[idx]);
+
+ return (sendAsyncResult64(reference64, result, args64, numArgs));
+}
+
+IOReturn IOUserClient::sendAsyncResult64WithOptions(OSAsyncReference64 reference,
+ IOReturn result, io_user_reference_t args[], UInt32 numArgs, IOOptionBits options)
+{
+ return _sendAsyncResult64(reference, result, args, numArgs, options);
+}
+
+IOReturn IOUserClient::sendAsyncResult64(OSAsyncReference64 reference,
+ IOReturn result, io_user_reference_t args[], UInt32 numArgs)
+{
+ return _sendAsyncResult64(reference, result, args, numArgs, 0);
+}
+
+IOReturn IOUserClient::_sendAsyncResult64(OSAsyncReference64 reference,
+ IOReturn result, io_user_reference_t args[], UInt32 numArgs, IOOptionBits options)
+{
+ struct ReplyMsg
+ {
+ mach_msg_header_t msgHdr;
+ union
+ {
+ struct
+ {
+ OSNotificationHeader notifyHdr;
+ IOAsyncCompletionContent asyncContent;
+ uint32_t args[kMaxAsyncArgs];
+ } msg32;
+ struct
+ {
+ OSNotificationHeader64 notifyHdr;
+ IOAsyncCompletionContent asyncContent;
+ io_user_reference_t args[kMaxAsyncArgs] __attribute__ ((packed));
+ } msg64;
+ } m;
};
- ReplyMsg replyMsg;
- mach_port_t replyPort;
+ ReplyMsg replyMsg;
+ mach_port_t replyPort;
kern_return_t kr;
// If no reply port, do nothing.
- replyPort = (mach_port_t) reference[0];
- if(replyPort == MACH_PORT_NULL)
+ replyPort = (mach_port_t) (reference[0] & ~kIOUCAsync0Flags);
+ if (replyPort == MACH_PORT_NULL)
return kIOReturnSuccess;
- if(numArgs > kMaxAsyncArgs)
+ if (numArgs > kMaxAsyncArgs)
return kIOReturnMessageTooLarge;
+
replyMsg.msgHdr.msgh_bits = MACH_MSGH_BITS(MACH_MSG_TYPE_COPY_SEND /*remote*/,
- 0 /*local*/);
- replyMsg.msgHdr.msgh_size =
- sizeof(replyMsg) - (kMaxAsyncArgs-numArgs)*sizeof(void *);
+ 0 /*local*/);
replyMsg.msgHdr.msgh_remote_port = replyPort;
- replyMsg.msgHdr.msgh_local_port = 0;
- replyMsg.msgHdr.msgh_id = kOSNotificationMessageID;
-
- replyMsg.notifyHdr.size = sizeof(IOAsyncCompletionContent)
- + numArgs*sizeof(void *);
- replyMsg.notifyHdr.type = kIOAsyncCompletionNotificationType;
- bcopy( reference, replyMsg.notifyHdr.reference, sizeof(OSAsyncReference));
-
- replyMsg.asyncContent.result = result;
- if(numArgs > 0)
- bcopy(args, replyMsg.args, sizeof(void *)*numArgs);
- kr = mach_msg_send_from_kernel( &replyMsg.msgHdr,
- replyMsg.msgHdr.msgh_size);
- if( KERN_SUCCESS != kr)
- IOLog("%s: mach_msg_send_from_kernel {%x}\n", __FILE__, kr );
+ replyMsg.msgHdr.msgh_local_port = 0;
+ replyMsg.msgHdr.msgh_id = kOSNotificationMessageID;
+ if (kIOUCAsync64Flag & reference[0])
+ {
+ replyMsg.msgHdr.msgh_size =
+ sizeof(replyMsg.msgHdr) + sizeof(replyMsg.m.msg64)
+ - (kMaxAsyncArgs - numArgs) * sizeof(io_user_reference_t);
+ replyMsg.m.msg64.notifyHdr.size = sizeof(IOAsyncCompletionContent)
+ + numArgs * sizeof(io_user_reference_t);
+ replyMsg.m.msg64.notifyHdr.type = kIOAsyncCompletionNotificationType;
+ bcopy(reference, replyMsg.m.msg64.notifyHdr.reference, sizeof(OSAsyncReference64));
+
+ replyMsg.m.msg64.asyncContent.result = result;
+ if (numArgs)
+ bcopy(args, replyMsg.m.msg64.args, numArgs * sizeof(io_user_reference_t));
+ }
+ else
+ {
+ unsigned int idx;
+
+ replyMsg.msgHdr.msgh_size =
+ sizeof(replyMsg.msgHdr) + sizeof(replyMsg.m.msg32)
+ - (kMaxAsyncArgs - numArgs) * sizeof(uint32_t);
+
+ replyMsg.m.msg32.notifyHdr.size = sizeof(IOAsyncCompletionContent)
+ + numArgs * sizeof(uint32_t);
+ replyMsg.m.msg32.notifyHdr.type = kIOAsyncCompletionNotificationType;
+
+ for (idx = 0; idx < kOSAsyncRefCount; idx++)
+ replyMsg.m.msg32.notifyHdr.reference[idx] = REF32(reference[idx]);
+
+ replyMsg.m.msg32.asyncContent.result = result;
+
+ for (idx = 0; idx < numArgs; idx++)
+ replyMsg.m.msg32.args[idx] = REF32(args[idx]);
+ }
+
+ if ((options & kIOUserNotifyOptionCanDrop) != 0) {
+ kr = mach_msg_send_from_kernel_with_options( &replyMsg.msgHdr,
+ replyMsg.msgHdr.msgh_size, MACH_SEND_TIMEOUT, MACH_MSG_TIMEOUT_NONE);
+ } else {
+ /* Fail on full queue. */
+ kr = mach_msg_send_from_kernel_proper( &replyMsg.msgHdr,
+ replyMsg.msgHdr.msgh_size);
+ }
+ if ((KERN_SUCCESS != kr) && (MACH_SEND_TIMED_OUT != kr))
+ IOLog("%s: mach_msg_send_from_kernel_proper {%x}\n", __FILE__, kr );
return kr;
}
+
/* * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * */
extern "C" {
if( !(out = OSDynamicCast( cls, obj))) \
return( kIOReturnBadArgument )
+/* * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * */
+
+/* Routine io_server_version */
+kern_return_t is_io_server_version(
+ mach_port_t master_port,
+ uint64_t *version)
+{
+ *version = IOKIT_SERVER_VERSION;
+ return (kIOReturnSuccess);
+}
+
/* Routine io_object_get_class */
kern_return_t is_io_object_get_class(
io_object_t object,
io_name_t className )
{
const OSMetaClass* my_obj = NULL;
-
- if( !object)
- return( kIOReturnBadArgument );
+
+ if( !object)
+ return( kIOReturnBadArgument );
my_obj = object->getMetaClass();
if (!my_obj) {
return (kIOReturnNotFound);
}
- strcpy( className, my_obj->getClassName());
+ strlcpy( className, my_obj->getClassName(), sizeof(io_name_t));
return( kIOReturnSuccess );
}
my_cstr = superclass->getClassName();
if (my_cstr) {
- strncpy(class_name, my_cstr, sizeof(io_name_t)-1);
+ strlcpy(class_name, my_cstr, sizeof(io_name_t));
return( kIOReturnSuccess );
}
return (kIOReturnNotFound);
my_cstr = identifier->getCStringNoCopy();
if (my_cstr) {
- strncpy(bundle_name, identifier->getCStringNoCopy(), sizeof(io_name_t)-1);
+ strlcpy(bundle_name, identifier->getCStringNoCopy(), sizeof(io_name_t));
return( kIOReturnSuccess );
}
/* Routine io_object_get_retain_count */
kern_return_t is_io_object_get_retain_count(
io_object_t object,
- int *retainCount )
+ uint32_t *retainCount )
{
if( !object)
return( kIOReturnBadArgument );
return( kIOReturnSuccess );
}
-/* Routine io_service_match_property_table */
-kern_return_t is_io_service_match_property_table(
+
+static kern_return_t internal_io_service_match_property_table(
io_service_t _service,
- io_string_t matching,
- boolean_t *matches )
+ const char * matching,
+ mach_msg_type_number_t matching_size,
+ boolean_t *matches)
{
CHECK( IOService, _service, service );
OSObject * obj;
OSDictionary * dict;
- obj = OSUnserializeXML( matching );
-
+ obj = matching_size ? OSUnserializeXML(matching, matching_size)
+ : OSUnserializeXML(matching);
if( (dict = OSDynamicCast( OSDictionary, obj))) {
*matches = service->passiveMatch( dict );
kr = kIOReturnSuccess;
return( kr );
}
+/* Routine io_service_match_property_table */
+kern_return_t is_io_service_match_property_table(
+ io_service_t service,
+ io_string_t matching,
+ boolean_t *matches )
+{
+ return (internal_io_service_match_property_table(service, matching, 0, matches));
+}
+
+
/* Routine io_service_match_property_table_ool */
kern_return_t is_io_service_match_property_table_ool(
io_object_t service,
io_buf_ptr_t matching,
mach_msg_type_number_t matchingCnt,
- natural_t *result,
+ kern_return_t *result,
boolean_t *matches )
{
- kern_return_t kr;
- vm_offset_t data;
- vm_map_offset_t map_data;
+ kern_return_t kr;
+ vm_offset_t data;
+ vm_map_offset_t map_data;
kr = vm_map_copyout( kernel_map, &map_data, (vm_map_copy_t) matching );
data = CAST_DOWN(vm_offset_t, map_data);
if( KERN_SUCCESS == kr) {
// must return success after vm_map_copyout() succeeds
- *result = is_io_service_match_property_table( service,
- (char *) data, matches );
+ *result = internal_io_service_match_property_table(service,
+ (const char *)data, matchingCnt, matches );
vm_deallocate( kernel_map, data, matchingCnt );
}
return( kr );
}
-/* Routine io_service_get_matching_services */
-kern_return_t is_io_service_get_matching_services(
+/* Routine io_service_match_property_table_bin */
+kern_return_t is_io_service_match_property_table_bin(
+ io_object_t service,
+ io_struct_inband_t matching,
+ mach_msg_type_number_t matchingCnt,
+ boolean_t *matches)
+{
+ return (internal_io_service_match_property_table(service, matching, matchingCnt, matches));
+}
+
+static kern_return_t internal_io_service_get_matching_services(
mach_port_t master_port,
- io_string_t matching,
+ const char * matching,
+ mach_msg_type_number_t matching_size,
io_iterator_t *existing )
{
kern_return_t kr;
if( master_port != master_device_port)
return( kIOReturnNotPrivileged);
- obj = OSUnserializeXML( matching );
-
+ obj = matching_size ? OSUnserializeXML(matching, matching_size)
+ : OSUnserializeXML(matching);
if( (dict = OSDynamicCast( OSDictionary, obj))) {
*existing = IOService::getMatchingServices( dict );
kr = kIOReturnSuccess;
return( kr );
}
+/* Routine io_service_get_matching_services */
+kern_return_t is_io_service_get_matching_services(
+ mach_port_t master_port,
+ io_string_t matching,
+ io_iterator_t *existing )
+{
+ return (internal_io_service_get_matching_services(master_port, matching, 0, existing));
+}
+
/* Routine io_service_get_matching_services_ool */
kern_return_t is_io_service_get_matching_services_ool(
mach_port_t master_port,
io_buf_ptr_t matching,
mach_msg_type_number_t matchingCnt,
- natural_t *result,
+ kern_return_t *result,
io_object_t *existing )
{
kern_return_t kr;
if( KERN_SUCCESS == kr) {
// must return success after vm_map_copyout() succeeds
- *result = is_io_service_get_matching_services( master_port,
- (char *) data, existing );
+ *result = internal_io_service_get_matching_services(master_port,
+ (const char *) data, matchingCnt, existing);
vm_deallocate( kernel_map, data, matchingCnt );
}
return( kr );
}
-/* Routine io_service_add_notification */
-kern_return_t is_io_service_add_notification(
+/* Routine io_service_get_matching_services_bin */
+kern_return_t is_io_service_get_matching_services_bin(
mach_port_t master_port,
- io_name_t notification_type,
- io_string_t matching,
- mach_port_t port,
- io_async_ref_t reference,
- mach_msg_type_number_t referenceCnt,
- io_object_t * notification )
+ io_struct_inband_t matching,
+ mach_msg_type_number_t matchingCnt,
+ io_object_t *existing)
{
- IOServiceUserNotification * userNotify = 0;
- IONotifier * notify = 0;
- const OSSymbol * sym;
- OSDictionary * dict;
- IOReturn err;
- unsigned long int userMsgType;
+ return (internal_io_service_get_matching_services(master_port, matching, matchingCnt, existing));
+}
+static kern_return_t internal_io_service_get_matching_service(
+ mach_port_t master_port,
+ const char * matching,
+ mach_msg_type_number_t matching_size,
+ io_service_t *service )
+{
+ kern_return_t kr;
+ OSObject * obj;
+ OSDictionary * dict;
+
if( master_port != master_device_port)
return( kIOReturnNotPrivileged);
- do {
- err = kIOReturnNoResources;
+ obj = matching_size ? OSUnserializeXML(matching, matching_size)
+ : OSUnserializeXML(matching);
+ if( (dict = OSDynamicCast( OSDictionary, obj))) {
+ *service = IOService::copyMatchingService( dict );
+ kr = *service ? kIOReturnSuccess : kIOReturnNotFound;
+ } else
+ kr = kIOReturnBadArgument;
- if( !(sym = OSSymbol::withCString( notification_type )))
- err = kIOReturnNoResources;
+ if( obj)
+ obj->release();
- if( !(dict = OSDynamicCast( OSDictionary,
- OSUnserializeXML( matching )))) {
- err = kIOReturnBadArgument;
- continue;
- }
+ return( kr );
+}
+
+/* Routine io_service_get_matching_service */
+kern_return_t is_io_service_get_matching_service(
+ mach_port_t master_port,
+ io_string_t matching,
+ io_service_t *service )
+{
+ return (internal_io_service_get_matching_service(master_port, matching, 0, service));
+}
+
+/* Routine io_service_get_matching_services_ool */
+kern_return_t is_io_service_get_matching_service_ool(
+ mach_port_t master_port,
+ io_buf_ptr_t matching,
+ mach_msg_type_number_t matchingCnt,
+ kern_return_t *result,
+ io_object_t *service )
+{
+ kern_return_t kr;
+ vm_offset_t data;
+ vm_map_offset_t map_data;
+
+ kr = vm_map_copyout( kernel_map, &map_data, (vm_map_copy_t) matching );
+ data = CAST_DOWN(vm_offset_t, map_data);
+
+ if( KERN_SUCCESS == kr) {
+ // must return success after vm_map_copyout() succeeds
+ *result = internal_io_service_get_matching_service(master_port,
+ (const char *) data, matchingCnt, service );
+ vm_deallocate( kernel_map, data, matchingCnt );
+ }
+
+ return( kr );
+}
+
+/* Routine io_service_get_matching_service_bin */
+kern_return_t is_io_service_get_matching_service_bin(
+ mach_port_t master_port,
+ io_struct_inband_t matching,
+ mach_msg_type_number_t matchingCnt,
+ io_object_t *service)
+{
+ return (internal_io_service_get_matching_service(master_port, matching, matchingCnt, service));
+}
+
+static kern_return_t internal_io_service_add_notification(
+ mach_port_t master_port,
+ io_name_t notification_type,
+ const char * matching,
+ size_t matching_size,
+ mach_port_t port,
+ void * reference,
+ vm_size_t referenceSize,
+ bool client64,
+ io_object_t * notification )
+{
+ IOServiceUserNotification * userNotify = 0;
+ IONotifier * notify = 0;
+ const OSSymbol * sym;
+ OSDictionary * dict;
+ IOReturn err;
+ unsigned long int userMsgType;
+
+ if( master_port != master_device_port)
+ return( kIOReturnNotPrivileged);
+
+ do {
+ err = kIOReturnNoResources;
+
+ if( !(sym = OSSymbol::withCString( notification_type )))
+ err = kIOReturnNoResources;
+
+ if (matching_size)
+ {
+ dict = OSDynamicCast(OSDictionary, OSUnserializeXML(matching, matching_size));
+ }
+ else
+ {
+ dict = OSDynamicCast(OSDictionary, OSUnserializeXML(matching));
+ }
+
+ if (!dict) {
+ err = kIOReturnBadArgument;
+ continue;
+ }
if( (sym == gIOPublishNotification)
|| (sym == gIOFirstPublishNotification))
userNotify = new IOServiceUserNotification;
if( userNotify && !userNotify->init( port, userMsgType,
- reference)) {
+ reference, referenceSize, client64)) {
+ iokit_release_port_send(port);
userNotify->release();
userNotify = 0;
}
if( !userNotify)
continue;
- notify = IOService::addNotification( sym, dict,
+ notify = IOService::addMatchingNotification( sym, dict,
&userNotify->_handler, userNotify );
if( notify) {
- dict = 0;
*notification = userNotify;
userNotify->setNotification( notify );
err = kIOReturnSuccess;
return( err );
}
-/* Routine io_service_add_notification_ool */
-kern_return_t is_io_service_add_notification_ool(
+
+/* Routine io_service_add_notification */
+kern_return_t is_io_service_add_notification(
mach_port_t master_port,
io_name_t notification_type,
- io_buf_ptr_t matching,
+ io_string_t matching,
+ mach_port_t port,
+ io_async_ref_t reference,
+ mach_msg_type_number_t referenceCnt,
+ io_object_t * notification )
+{
+ return (internal_io_service_add_notification(master_port, notification_type,
+ matching, 0, port, &reference[0], sizeof(io_async_ref_t),
+ false, notification));
+}
+
+/* Routine io_service_add_notification_64 */
+kern_return_t is_io_service_add_notification_64(
+ mach_port_t master_port,
+ io_name_t notification_type,
+ io_string_t matching,
+ mach_port_t wake_port,
+ io_async_ref64_t reference,
+ mach_msg_type_number_t referenceCnt,
+ io_object_t *notification )
+{
+ return (internal_io_service_add_notification(master_port, notification_type,
+ matching, 0, wake_port, &reference[0], sizeof(io_async_ref64_t),
+ true, notification));
+}
+
+/* Routine io_service_add_notification_bin */
+kern_return_t is_io_service_add_notification_bin
+(
+ mach_port_t master_port,
+ io_name_t notification_type,
+ io_struct_inband_t matching,
mach_msg_type_number_t matchingCnt,
mach_port_t wake_port,
io_async_ref_t reference,
mach_msg_type_number_t referenceCnt,
- natural_t *result,
+ io_object_t *notification)
+{
+ return (internal_io_service_add_notification(master_port, notification_type,
+ matching, matchingCnt, wake_port, &reference[0], sizeof(io_async_ref_t),
+ false, notification));
+}
+
+/* Routine io_service_add_notification_bin_64 */
+kern_return_t is_io_service_add_notification_bin_64
+(
+ mach_port_t master_port,
+ io_name_t notification_type,
+ io_struct_inband_t matching,
+ mach_msg_type_number_t matchingCnt,
+ mach_port_t wake_port,
+ io_async_ref64_t reference,
+ mach_msg_type_number_t referenceCnt,
+ io_object_t *notification)
+{
+ return (internal_io_service_add_notification(master_port, notification_type,
+ matching, matchingCnt, wake_port, &reference[0], sizeof(io_async_ref64_t),
+ true, notification));
+}
+
+static kern_return_t internal_io_service_add_notification_ool(
+ mach_port_t master_port,
+ io_name_t notification_type,
+ io_buf_ptr_t matching,
+ mach_msg_type_number_t matchingCnt,
+ mach_port_t wake_port,
+ void * reference,
+ vm_size_t referenceSize,
+ bool client64,
+ kern_return_t *result,
io_object_t *notification )
{
kern_return_t kr;
if( KERN_SUCCESS == kr) {
// must return success after vm_map_copyout() succeeds
- *result = is_io_service_add_notification( master_port, notification_type,
- (char *) data, wake_port, reference, referenceCnt, notification );
+ *result = internal_io_service_add_notification( master_port, notification_type,
+ (char *) data, matchingCnt, wake_port, reference, referenceSize, client64, notification );
vm_deallocate( kernel_map, data, matchingCnt );
}
return( kr );
}
+/* Routine io_service_add_notification_ool */
+kern_return_t is_io_service_add_notification_ool(
+ mach_port_t master_port,
+ io_name_t notification_type,
+ io_buf_ptr_t matching,
+ mach_msg_type_number_t matchingCnt,
+ mach_port_t wake_port,
+ io_async_ref_t reference,
+ mach_msg_type_number_t referenceCnt,
+ kern_return_t *result,
+ io_object_t *notification )
+{
+ return (internal_io_service_add_notification_ool(master_port, notification_type,
+ matching, matchingCnt, wake_port, &reference[0], sizeof(io_async_ref_t),
+ false, result, notification));
+}
+
+/* Routine io_service_add_notification_ool_64 */
+kern_return_t is_io_service_add_notification_ool_64(
+ mach_port_t master_port,
+ io_name_t notification_type,
+ io_buf_ptr_t matching,
+ mach_msg_type_number_t matchingCnt,
+ mach_port_t wake_port,
+ io_async_ref64_t reference,
+ mach_msg_type_number_t referenceCnt,
+ kern_return_t *result,
+ io_object_t *notification )
+{
+ return (internal_io_service_add_notification_ool(master_port, notification_type,
+ matching, matchingCnt, wake_port, &reference[0], sizeof(io_async_ref64_t),
+ true, result, notification));
+}
/* Routine io_service_add_notification_old */
kern_return_t is_io_service_add_notification_old(
io_name_t notification_type,
io_string_t matching,
mach_port_t port,
+ // for binary compatibility reasons, this must be natural_t for ILP32
natural_t ref,
io_object_t * notification )
{
matching, port, &ref, 1, notification ));
}
-/* Routine io_service_add_message_notification */
-kern_return_t is_io_service_add_interest_notification(
+
+static kern_return_t internal_io_service_add_interest_notification(
io_object_t _service,
io_name_t type_of_interest,
mach_port_t port,
- io_async_ref_t reference,
- mach_msg_type_number_t referenceCnt,
+ void * reference,
+ vm_size_t referenceSize,
+ bool client64,
io_object_t * notification )
{
userNotify = new IOServiceMessageUserNotification;
if( userNotify && !userNotify->init( port, kIOServiceMessageNotificationType,
- reference, kIOUserNotifyMaxMessageSize )) {
+ reference, referenceSize,
+ kIOUserNotifyMaxMessageSize,
+ client64 )) {
+ iokit_release_port_send(port);
userNotify->release();
userNotify = 0;
}
return( err );
}
+/* Routine io_service_add_message_notification */
+kern_return_t is_io_service_add_interest_notification(
+ io_object_t service,
+ io_name_t type_of_interest,
+ mach_port_t port,
+ io_async_ref_t reference,
+ mach_msg_type_number_t referenceCnt,
+ io_object_t * notification )
+{
+ return (internal_io_service_add_interest_notification(service, type_of_interest,
+ port, &reference[0], sizeof(io_async_ref_t), false, notification));
+}
+
+/* Routine io_service_add_interest_notification_64 */
+kern_return_t is_io_service_add_interest_notification_64(
+ io_object_t service,
+ io_name_t type_of_interest,
+ mach_port_t wake_port,
+ io_async_ref64_t reference,
+ mach_msg_type_number_t referenceCnt,
+ io_object_t *notification )
+{
+ return (internal_io_service_add_interest_notification(service, type_of_interest,
+ wake_port, &reference[0], sizeof(io_async_ref64_t), true, notification));
+}
+
+
/* Routine io_service_acknowledge_notification */
kern_return_t is_io_service_acknowledge_notification(
io_object_t _service,
{
CHECK( IOService, _service, service );
- return( service->acknowledgeNotification( (IONotificationRef) notify_ref,
+ return( service->acknowledgeNotification( (IONotificationRef)(uintptr_t) notify_ref,
(IOOptionBits) response ));
}
{
CHECK( IOUserClient, connection, client );
+ IOStatisticsClientCall();
return( client->getNotificationSemaphore( (UInt32) notification_type,
semaphore ));
}
kern_return_t is_io_registry_create_iterator(
mach_port_t master_port,
io_name_t plane,
- int options,
+ uint32_t options,
io_object_t *iterator )
{
if( master_port != master_device_port)
kern_return_t is_io_registry_entry_create_iterator(
io_object_t registry_entry,
io_name_t plane,
- int options,
+ uint32_t options,
io_object_t *iterator )
{
CHECK( IORegistryEntry, registry_entry, entry );
return( kIOReturnNotFound );
}
+/* Routine io_registry_entry_get_registry_entry_id */
+kern_return_t is_io_registry_entry_get_registry_entry_id(
+ io_object_t registry_entry,
+ uint64_t *entry_id )
+{
+ CHECK( IORegistryEntry, registry_entry, entry );
+
+ *entry_id = entry->getRegistryEntryID();
+
+ return (kIOReturnSuccess);
+}
+
// Create a vm_map_copy_t or kalloc'ed data for memory
// to be copied out. ipc will free after the copyout.
-static kern_return_t copyoutkdata( void * data, vm_size_t len,
+static kern_return_t copyoutkdata( const void * data, vm_size_t len,
io_buf_ptr_t * buf )
{
kern_return_t err;
kern_return_t is_io_registry_entry_get_property_bytes(
io_object_t registry_entry,
io_name_t property_name,
- io_scalar_inband_t buf,
+ io_struct_inband_t buf,
mach_msg_type_number_t *dataCnt )
{
OSObject * obj;
CHECK( IORegistryEntry, registry_entry, entry );
+#if CONFIG_MACF
+ if (0 != mac_iokit_check_get_property(kauth_cred_get(), entry, property_name))
+ return kIOReturnNotPermitted;
+#endif
+
obj = entry->copyProperty(property_name);
if( !obj)
return( kIOReturnNoResources );
CHECK( IORegistryEntry, registry_entry, entry );
+#if CONFIG_MACF
+ if (0 != mac_iokit_check_get_property(kauth_cred_get(), entry, property_name))
+ return kIOReturnNotPermitted;
+#endif
+
obj = entry->copyProperty(property_name);
if( !obj)
return( kIOReturnNotFound );
obj->release();
return( kIOReturnNoMemory );
}
- s->clearText();
if( obj->serialize( s )) {
len = s->getLength();
io_object_t registry_entry,
io_name_t plane,
io_name_t property_name,
- int options,
+ uint32_t options,
io_buf_ptr_t *properties,
mach_msg_type_number_t *propertiesCnt )
{
CHECK( IORegistryEntry, registry_entry, entry );
+#if CONFIG_MACF
+ if (0 != mac_iokit_check_get_property(kauth_cred_get(), entry, property_name))
+ return kIOReturnNotPermitted;
+#endif
+
obj = entry->copyProperty( property_name,
IORegistryEntry::getPlane( plane ), options);
if( !obj)
return( kIOReturnNoMemory );
}
- s->clearText();
-
if( obj->serialize( s )) {
len = s->getLength();
*propertiesCnt = len;
return( err );
}
+#if CONFIG_MACF
+
+static kern_return_t
+filteredProperties(IORegistryEntry *entry, OSDictionary *properties, OSDictionary **filteredp)
+{
+ kern_return_t err = 0;
+ OSDictionary *filtered = NULL;
+ OSCollectionIterator *iter = NULL;
+ OSSymbol *key;
+ OSObject *p;
+ kauth_cred_t cred = kauth_cred_get();
+
+ if (properties == NULL)
+ return kIOReturnUnsupported;
+
+ if ((iter = OSCollectionIterator::withCollection(properties)) == NULL ||
+ (filtered = OSDictionary::withCapacity(properties->getCapacity())) == NULL) {
+ err = kIOReturnNoMemory;
+ goto out;
+ }
+
+ while ((p = iter->getNextObject()) != NULL) {
+ if ((key = OSDynamicCast(OSSymbol, p)) == NULL ||
+ mac_iokit_check_get_property(cred, entry, key->getCStringNoCopy()) != 0)
+ continue;
+ filtered->setObject(key, properties->getObject(key));
+ }
+
+out:
+ if (iter != NULL)
+ iter->release();
+ *filteredp = filtered;
+ return err;
+}
+
+#endif
+
/* Routine io_registry_entry_get_properties */
kern_return_t is_io_registry_entry_get_properties(
io_object_t registry_entry,
io_buf_ptr_t *properties,
mach_msg_type_number_t *propertiesCnt )
{
- kern_return_t err;
+ kern_return_t err = 0;
vm_size_t len;
CHECK( IORegistryEntry, registry_entry, entry );
if( !s)
return( kIOReturnNoMemory );
- s->clearText();
+ if (!entry->serializeProperties(s))
+ err = kIOReturnUnsupported;
+
+#if CONFIG_MACF
+ if (!err && mac_iokit_check_filter_properties(kauth_cred_get(), entry)) {
+ OSObject *propobj = OSUnserializeXML(s->text(), s->getLength());
+ OSDictionary *filteredprops = NULL;
+ err = filteredProperties(entry, OSDynamicCast(OSDictionary, propobj), &filteredprops);
+ if (propobj) propobj->release();
+
+ if (!err) {
+ s->clearText();
+ if (!filteredprops->serialize(s))
+ err = kIOReturnUnsupported;
+ }
+ if (filteredprops != NULL)
+ filteredprops->release();
+ }
+#endif /* CONFIG_MACF */
+
+ if (!err) {
+ len = s->getLength();
+ *propertiesCnt = len;
+ err = copyoutkdata( s->text(), len, properties );
+ }
+
+ s->release();
+ return( err );
+}
+
+#if CONFIG_MACF
+
+struct GetPropertiesEditorRef
+{
+ kauth_cred_t cred;
+ IORegistryEntry * entry;
+ OSCollection * root;
+};
+
+static const OSMetaClassBase *
+GetPropertiesEditor(void * reference,
+ OSSerialize * s,
+ OSCollection * container,
+ const OSSymbol * name,
+ const OSMetaClassBase * value)
+{
+ GetPropertiesEditorRef * ref = (typeof(ref)) reference;
+
+ if (!ref->root) ref->root = container;
+ if (ref->root == container)
+ {
+ if (0 != mac_iokit_check_get_property(ref->cred, ref->entry, name->getCStringNoCopy()))
+ {
+ value = 0;
+ }
+ }
+ if (value) value->retain();
+ return (value);
+}
+
+#endif /* CONFIG_MACF */
+
+/* Routine io_registry_entry_get_properties */
+kern_return_t is_io_registry_entry_get_properties_bin(
+ io_object_t registry_entry,
+ io_buf_ptr_t *properties,
+ mach_msg_type_number_t *propertiesCnt)
+{
+ kern_return_t err = kIOReturnSuccess;
+ vm_size_t len;
+ OSSerialize * s;
+ OSSerialize::Editor editor = 0;
+ void * editRef = 0;
+
+ CHECK(IORegistryEntry, registry_entry, entry);
+
+#if CONFIG_MACF
+ GetPropertiesEditorRef ref;
+ if (mac_iokit_check_filter_properties(kauth_cred_get(), entry))
+ {
+ editor = &GetPropertiesEditor;
+ editRef = &ref;
+ ref.cred = kauth_cred_get();
+ ref.entry = entry;
+ ref.root = 0;
+ }
+#endif
+
+ s = OSSerialize::binaryWithCapacity(4096, editor, editRef);
+ if (!s) return (kIOReturnNoMemory);
+
+ if (!entry->serializeProperties(s)) err = kIOReturnUnsupported;
+
+ if (kIOReturnSuccess == err)
+ {
+ len = s->getLength();
+ *propertiesCnt = len;
+ err = copyoutkdata(s->text(), len, properties);
+ }
+ s->release();
+
+ return (err);
+}
+
+/* Routine io_registry_entry_get_property_bin */
+kern_return_t is_io_registry_entry_get_property_bin(
+ io_object_t registry_entry,
+ io_name_t plane,
+ io_name_t property_name,
+ uint32_t options,
+ io_buf_ptr_t *properties,
+ mach_msg_type_number_t *propertiesCnt )
+{
+ kern_return_t err;
+ vm_size_t len;
+ OSObject * obj;
+ const OSSymbol * sym;
+
+ CHECK( IORegistryEntry, registry_entry, entry );
+
+#if CONFIG_MACF
+ if (0 != mac_iokit_check_get_property(kauth_cred_get(), entry, property_name))
+ return kIOReturnNotPermitted;
+#endif
+
+ if ((kIORegistryIterateRecursively & options) && plane[0])
+ {
+ obj = entry->copyProperty(property_name,
+ IORegistryEntry::getPlane(plane), options);
+ }
+ else
+ {
+ obj = entry->copyProperty(property_name);
+ }
+
+ if( !obj)
+ return( kIOReturnNotFound );
+
+ sym = OSSymbol::withCString(property_name);
+ if (sym)
+ {
+ if (gIORemoveOnReadProperties->containsObject(sym)) entry->removeProperty(sym);
+ sym->release();
+ }
+
+ OSSerialize * s = OSSerialize::binaryWithCapacity(4096);
+ if( !s) {
+ obj->release();
+ return( kIOReturnNoMemory );
+ }
- if( entry->serializeProperties( s )) {
+ if( obj->serialize( s )) {
len = s->getLength();
*propertiesCnt = len;
err = copyoutkdata( s->text(), len, properties );
- } else
- err = kIOReturnUnsupported;
+ } else err = kIOReturnUnsupported;
s->release();
+ obj->release();
return( err );
}
io_object_t registry_entry,
io_buf_ptr_t properties,
mach_msg_type_number_t propertiesCnt,
- natural_t * result)
+ kern_return_t * result)
{
OSObject * obj;
kern_return_t err;
CHECK( IORegistryEntry, registry_entry, entry );
+ if( propertiesCnt > sizeof(io_struct_inband_t) * 1024)
+ return( kIOReturnMessageTooLarge);
+
err = vm_map_copyout( kernel_map, &map_data, (vm_map_copy_t) properties );
data = CAST_DOWN(vm_offset_t, map_data);
if( KERN_SUCCESS == err) {
// must return success after vm_map_copyout() succeeds
- obj = OSUnserializeXML( (const char *) data );
+ obj = OSUnserializeXML( (const char *) data, propertiesCnt );
vm_deallocate( kernel_map, data, propertiesCnt );
- if( obj) {
- res = entry->setProperties( obj );
- obj->release();
- } else
- res = kIOReturnBadArgument;
- } else
+ if (!obj)
+ res = kIOReturnBadArgument;
+#if CONFIG_MACF
+ else if (0 != mac_iokit_check_set_properties(kauth_cred_get(),
+ registry_entry, obj))
+ {
+ res = kIOReturnNotPermitted;
+ }
+#endif
+ else
+ {
+ res = entry->setProperties( obj );
+ }
+
+ if (obj)
+ obj->release();
+ } else
res = err;
*result = res;
/* Routine io_service_get_busy_state */
kern_return_t is_io_service_get_busy_state(
io_object_t _service,
- int *busyState )
+ uint32_t *busyState )
{
CHECK( IOService, _service, service );
/* Routine io_service_get_state */
kern_return_t is_io_service_get_state(
io_object_t _service,
- uint64_t *state )
+ uint64_t *state,
+ uint32_t *busy_state,
+ uint64_t *accumulated_busy_time )
{
CHECK( IOService, _service, service );
- *state = service->getState();
+ *state = service->getState();
+ *busy_state = service->getBusyState();
+ *accumulated_busy_time = service->getAccumulatedBusyTime();
return( kIOReturnSuccess );
}
io_object_t _service,
mach_timespec_t wait_time )
{
+ uint64_t timeoutNS;
+
CHECK( IOService, _service, service );
- return( service->waitQuiet( &wait_time ));
+ timeoutNS = wait_time.tv_sec;
+ timeoutNS *= kSecondScale;
+ timeoutNS += wait_time.tv_nsec;
+
+ return( service->waitQuiet(timeoutNS) );
}
/* Routine io_service_request_probe */
kern_return_t is_io_service_request_probe(
io_object_t _service,
- int options )
+ uint32_t options )
{
CHECK( IOService, _service, service );
return( service->requestProbe( options ));
}
-
-/* Routine io_service_open */
-kern_return_t is_io_service_open(
+/* Routine io_service_get_authorization_id */
+kern_return_t is_io_service_get_authorization_id(
io_object_t _service,
- task_t owningTask,
- int connect_type,
- io_object_t *connection )
+ uint64_t *authorization_id )
{
- IOUserClient * client;
- IOReturn err;
+ kern_return_t kr;
CHECK( IOService, _service, service );
- err = service->newUserClient( owningTask, (void *) owningTask,
- connect_type, 0, &client );
+ kr = IOUserClient::clientHasPrivilege( (void *) current_task(),
+ kIOClientPrivilegeAdministrator );
+ if( kIOReturnSuccess != kr)
+ return( kr );
+
+ *authorization_id = service->getAuthorizationID();
- if( err == kIOReturnSuccess) {
- assert( OSDynamicCast(IOUserClient, client) );
- *connection = client;
- }
+ return( kr );
+}
- return( err);
+/* Routine io_service_set_authorization_id */
+kern_return_t is_io_service_set_authorization_id(
+ io_object_t _service,
+ uint64_t authorization_id )
+{
+ CHECK( IOService, _service, service );
+
+ return( service->setAuthorizationID( authorization_id ) );
}
/* Routine io_service_open_ndr */
kern_return_t is_io_service_open_extended(
io_object_t _service,
task_t owningTask,
- int connect_type,
+ uint32_t connect_type,
NDR_record_t ndr,
io_buf_ptr_t properties,
mach_msg_type_number_t propertiesCnt,
- natural_t * result,
+ kern_return_t * result,
io_object_t *connection )
{
IOUserClient * client = 0;
vm_offset_t data;
vm_map_offset_t map_data;
+ if( propertiesCnt > sizeof(io_struct_inband_t))
+ return( kIOReturnMessageTooLarge);
+
err = vm_map_copyout( kernel_map, &map_data, (vm_map_copy_t) properties );
res = err;
data = CAST_DOWN(vm_offset_t, map_data);
if (KERN_SUCCESS == err)
{
// must return success after vm_map_copyout() succeeds
- obj = OSUnserializeXML( (const char *) data );
+ obj = OSUnserializeXML( (const char *) data, propertiesCnt );
vm_deallocate( kernel_map, data, propertiesCnt );
propertiesDict = OSDynamicCast(OSDictionary, obj);
if (!propertiesDict)
disallowAccess = (crossEndian
&& (kOSBooleanTrue != service->getProperty(kIOUserClientCrossEndianCompatibleKey))
&& (kOSBooleanTrue != client->getProperty(kIOUserClientCrossEndianCompatibleKey)));
-
- if (disallowAccess)
+ if (disallowAccess) res = kIOReturnUnsupported;
+#if CONFIG_MACF
+ else if (0 != mac_iokit_check_open(kauth_cred_get(), client, connect_type))
+ res = kIOReturnNotPermitted;
+#endif
+ if (kIOReturnSuccess != res)
{
+ IOStatisticsClientCall();
client->clientClose();
client->release();
client = 0;
- res = kIOReturnUnsupported;
break;
}
client->sharedInstance = (0 != client->getProperty(kIOUserClientSharedInstanceKey));
+ OSString * creatorName = IOCopyLogNameForPID(proc_selfpid());
+ if (creatorName)
+ {
+ client->setProperty(kIOUserClientCreatorKey, creatorName);
+ creatorName->release();
+ }
+ client->setTerminateDefer(service, false);
}
}
while (false);
CHECK( IOUserClient, connection, client );
+ IOStatisticsClientCall();
client->clientClose();
return( kIOReturnSuccess );
/* Routine io_connect_set_notification_port */
kern_return_t is_io_connect_set_notification_port(
io_object_t connection,
- int notification_type,
+ uint32_t notification_type,
+ mach_port_t port,
+ uint32_t reference)
+{
+ CHECK( IOUserClient, connection, client );
+
+ IOStatisticsClientCall();
+ return( client->registerNotificationPort( port, notification_type,
+ (io_user_reference_t) reference ));
+}
+
+/* Routine io_connect_set_notification_port */
+kern_return_t is_io_connect_set_notification_port_64(
+ io_object_t connection,
+ uint32_t notification_type,
mach_port_t port,
- int reference)
+ io_user_reference_t reference)
{
CHECK( IOUserClient, connection, client );
+ IOStatisticsClientCall();
return( client->registerNotificationPort( port, notification_type,
reference ));
}
-kern_return_t is_io_connect_map_memory(
- io_object_t connect,
- int type,
- task_t task,
- vm_address_t * mapAddr,
- vm_size_t * mapSize,
- int flags )
+/* Routine io_connect_map_memory_into_task */
+kern_return_t is_io_connect_map_memory_into_task
+(
+ io_connect_t connection,
+ uint32_t memory_type,
+ task_t into_task,
+ mach_vm_address_t *address,
+ mach_vm_size_t *size,
+ uint32_t flags
+)
{
IOReturn err;
IOMemoryMap * map;
- CHECK( IOUserClient, connect, client );
+ CHECK( IOUserClient, connection, client );
- map = client->mapClientMemory( type, task, flags, *mapAddr );
+ IOStatisticsClientCall();
+ map = client->mapClientMemory64( memory_type, into_task, flags, *address );
if( map) {
- *mapAddr = map->getVirtualAddress();
- if( mapSize)
- *mapSize = map->getLength();
+ *address = map->getAddress();
+ if( size)
+ *size = map->getSize();
if( client->sharedInstance
- || (task != current_task())) {
+ || (into_task != current_task())) {
// push a name out to the task owning the map,
// so we can clean up maps
-#if IOASSERT
- mach_port_name_t name =
-#endif
- IOMachPort::makeSendRightForTask(
- task, map, IKOT_IOKIT_OBJECT );
- assert( name );
+ mach_port_name_t name __unused =
+ IOMachPort::makeSendRightForTask(
+ into_task, map, IKOT_IOKIT_OBJECT );
} else {
// keep it with the user client
return( err );
}
+/* Routine is_io_connect_map_memory */
+kern_return_t is_io_connect_map_memory(
+ io_object_t connect,
+ uint32_t type,
+ task_t task,
+ uint32_t * mapAddr,
+ uint32_t * mapSize,
+ uint32_t flags )
+{
+ IOReturn err;
+ mach_vm_address_t address;
+ mach_vm_size_t size;
+
+ address = SCALAR64(*mapAddr);
+ size = SCALAR64(*mapSize);
+
+ err = is_io_connect_map_memory_into_task(connect, type, task, &address, &size, flags);
+
+ *mapAddr = SCALAR32(address);
+ *mapSize = SCALAR32(size);
+
+ return (err);
+}
+
+} /* extern "C" */
+
IOMemoryMap * IOUserClient::removeMappingForDescriptor(IOMemoryDescriptor * mem)
{
OSIterator * iter;
return (map);
}
-kern_return_t is_io_connect_unmap_memory(
- io_object_t connect,
- int type,
- task_t task,
- vm_address_t mapAddr )
+extern "C" {
+
+/* Routine io_connect_unmap_memory_from_task */
+kern_return_t is_io_connect_unmap_memory_from_task
+(
+ io_connect_t connection,
+ uint32_t memory_type,
+ task_t from_task,
+ mach_vm_address_t address)
{
IOReturn err;
IOOptionBits options = 0;
IOMemoryDescriptor * memory;
IOMemoryMap * map;
- CHECK( IOUserClient, connect, client );
+ CHECK( IOUserClient, connection, client );
- err = client->clientMemoryForType( (UInt32) type, &options, &memory );
+ IOStatisticsClientCall();
+ err = client->clientMemoryForType( (UInt32) memory_type, &options, &memory );
if( memory && (kIOReturnSuccess == err)) {
options = (options & ~kIOMapUserOptionsMask)
| kIOMapAnywhere | kIOMapReference;
- map = memory->map( task, mapAddr, options );
+ map = memory->createMappingInTask( from_task, address, options );
memory->release();
if( map)
{
IOLockUnlock( gIOObjectPortLock);
mach_port_name_t name = 0;
- if (task != current_task())
- name = IOMachPort::makeSendRightForTask( task, map, IKOT_IOKIT_OBJECT );
+ if (from_task != current_task())
+ name = IOMachPort::makeSendRightForTask( from_task, map, IKOT_IOKIT_OBJECT );
if (name)
{
- map->unmap();
- err = iokit_mod_send_right( task, name, -2 );
+ map->userClientUnmap();
+ err = iokit_mod_send_right( from_task, name, -2 );
err = kIOReturnSuccess;
}
else
IOMachPort::releasePortForObject( map, IKOT_IOKIT_OBJECT );
- if (task == current_task())
+ if (from_task == current_task())
map->release();
}
else
return( err );
}
+kern_return_t is_io_connect_unmap_memory(
+ io_object_t connect,
+ uint32_t type,
+ task_t task,
+ uint32_t mapAddr )
+{
+ IOReturn err;
+ mach_vm_address_t address;
+
+ address = SCALAR64(mapAddr);
+
+ err = is_io_connect_unmap_memory_from_task(connect, type, task, mapAddr);
+
+ return (err);
+}
+
/* Routine io_connect_add_client */
kern_return_t is_io_connect_add_client(
CHECK( IOUserClient, connection, client );
CHECK( IOUserClient, connect_to, to );
+ IOStatisticsClientCall();
return( client->connectClient( to ) );
}
io_object_t connection,
io_buf_ptr_t properties,
mach_msg_type_number_t propertiesCnt,
- natural_t * result)
+ kern_return_t * result)
{
return( is_io_registry_entry_set_properties( connection, properties, propertiesCnt, result ));
}
-
-/* Routine io_connect_method_scalarI_scalarO */
-kern_return_t is_io_connect_method_scalarI_scalarO(
- io_object_t connect,
- UInt32 index,
- void * input[],
- IOByteCount inputCount,
- void * output[],
- IOByteCount * outputCount )
+/* Routine io_user_client_method */
+kern_return_t is_io_connect_method_var_output
+(
+ io_connect_t connection,
+ uint32_t selector,
+ io_scalar_inband64_t scalar_input,
+ mach_msg_type_number_t scalar_inputCnt,
+ io_struct_inband_t inband_input,
+ mach_msg_type_number_t inband_inputCnt,
+ mach_vm_address_t ool_input,
+ mach_vm_size_t ool_input_size,
+ io_struct_inband_t inband_output,
+ mach_msg_type_number_t *inband_outputCnt,
+ io_scalar_inband64_t scalar_output,
+ mach_msg_type_number_t *scalar_outputCnt,
+ io_buf_ptr_t *var_output,
+ mach_msg_type_number_t *var_outputCnt
+)
{
- IOReturn err;
- IOExternalMethod * method;
- IOService * object;
- IOMethod func;
+ CHECK( IOUserClient, connection, client );
- CHECK( IOUserClient, connect, client);
- if( (method = client->getTargetAndMethodForIndex(&object, index))) {
- do {
- err = kIOReturnBadArgument;
- if( kIOUCScalarIScalarO != (method->flags & kIOUCTypeMask))
- continue;
- if( inputCount != method->count0)
- continue;
- if( *outputCount != method->count1)
- continue;
+ IOExternalMethodArguments args;
+ IOReturn ret;
+ IOMemoryDescriptor * inputMD = 0;
+ OSObject * structureVariableOutputData = 0;
- func = method->func;
+ bzero(&args.__reserved[0], sizeof(args.__reserved));
+ args.version = kIOExternalMethodArgumentsCurrentVersion;
- switch( inputCount) {
+ args.selector = selector;
- case 6:
- err = (object->*func)( input[0], input[1], input[2],
- input[3], input[4], input[5] );
- break;
- case 5:
- err = (object->*func)( input[0], input[1], input[2],
- input[3], input[4],
- &output[0] );
- break;
- case 4:
- err = (object->*func)( input[0], input[1], input[2],
- input[3],
- &output[0], &output[1] );
- break;
- case 3:
- err = (object->*func)( input[0], input[1], input[2],
- &output[0], &output[1], &output[2] );
- break;
- case 2:
- err = (object->*func)( input[0], input[1],
- &output[0], &output[1], &output[2],
- &output[3] );
- break;
- case 1:
- err = (object->*func)( input[0],
- &output[0], &output[1], &output[2],
- &output[3], &output[4] );
- break;
- case 0:
- err = (object->*func)( &output[0], &output[1], &output[2],
- &output[3], &output[4], &output[5] );
- break;
+ args.asyncWakePort = MACH_PORT_NULL;
+ args.asyncReference = 0;
+ args.asyncReferenceCount = 0;
+ args.structureVariableOutputData = &structureVariableOutputData;
- default:
- IOLog("%s: Bad method table\n", client->getName());
+ args.scalarInput = scalar_input;
+ args.scalarInputCount = scalar_inputCnt;
+ args.structureInput = inband_input;
+ args.structureInputSize = inband_inputCnt;
+
+ if (ool_input)
+ inputMD = IOMemoryDescriptor::withAddressRange(ool_input, ool_input_size,
+ kIODirectionOut, current_task());
+
+ args.structureInputDescriptor = inputMD;
+
+ args.scalarOutput = scalar_output;
+ args.scalarOutputCount = *scalar_outputCnt;
+ bzero(&scalar_output[0], *scalar_outputCnt * sizeof(scalar_output[0]));
+ args.structureOutput = inband_output;
+ args.structureOutputSize = *inband_outputCnt;
+ args.structureOutputDescriptor = NULL;
+ args.structureOutputDescriptorSize = 0;
+
+ IOStatisticsClientCall();
+ ret = client->externalMethod( selector, &args );
+
+ *scalar_outputCnt = args.scalarOutputCount;
+ *inband_outputCnt = args.structureOutputSize;
+
+ if (var_outputCnt && var_output && (kIOReturnSuccess == ret))
+ {
+ OSSerialize * serialize;
+ OSData * data;
+ vm_size_t len;
+
+ if ((serialize = OSDynamicCast(OSSerialize, structureVariableOutputData)))
+ {
+ len = serialize->getLength();
+ *var_outputCnt = len;
+ ret = copyoutkdata(serialize->text(), len, var_output);
+ }
+ else if ((data = OSDynamicCast(OSData, structureVariableOutputData)))
+ {
+ len = data->getLength();
+ *var_outputCnt = len;
+ ret = copyoutkdata(data->getBytesNoCopy(), len, var_output);
}
- } while( false);
+ else
+ {
+ ret = kIOReturnUnderrun;
+ }
+ }
- } else
- err = kIOReturnUnsupported;
+ if (inputMD)
+ inputMD->release();
+ if (structureVariableOutputData)
+ structureVariableOutputData->release();
- return( err);
+ return (ret);
}
-/* Routine io_connect_method_scalarI_structureO */
-kern_return_t is_io_connect_method_scalarI_structureO(
- io_object_t connect,
- UInt32 index,
- void * input[],
- IOByteCount inputCount,
- void * output,
- IOByteCount * outputCount )
+/* Routine io_user_client_method */
+kern_return_t is_io_connect_method
+(
+ io_connect_t connection,
+ uint32_t selector,
+ io_scalar_inband64_t scalar_input,
+ mach_msg_type_number_t scalar_inputCnt,
+ io_struct_inband_t inband_input,
+ mach_msg_type_number_t inband_inputCnt,
+ mach_vm_address_t ool_input,
+ mach_vm_size_t ool_input_size,
+ io_struct_inband_t inband_output,
+ mach_msg_type_number_t *inband_outputCnt,
+ io_scalar_inband64_t scalar_output,
+ mach_msg_type_number_t *scalar_outputCnt,
+ mach_vm_address_t ool_output,
+ mach_vm_size_t *ool_output_size
+)
{
- IOReturn err;
- IOExternalMethod * method;
- IOService * object;
- IOMethod func;
+ CHECK( IOUserClient, connection, client );
- CHECK( IOUserClient, connect, client);
+ IOExternalMethodArguments args;
+ IOReturn ret;
+ IOMemoryDescriptor * inputMD = 0;
+ IOMemoryDescriptor * outputMD = 0;
- if( (method = client->getTargetAndMethodForIndex(&object, index)) ) {
- do {
- err = kIOReturnBadArgument;
- if( kIOUCScalarIStructO != (method->flags & kIOUCTypeMask))
- continue;
- if( inputCount != method->count0)
- continue;
- if( (0xffffffff != method->count1)
- && (*outputCount != method->count1))
- continue;
+ bzero(&args.__reserved[0], sizeof(args.__reserved));
+ args.version = kIOExternalMethodArgumentsCurrentVersion;
- func = method->func;
+ args.selector = selector;
- switch( inputCount) {
+ args.asyncWakePort = MACH_PORT_NULL;
+ args.asyncReference = 0;
+ args.asyncReferenceCount = 0;
+ args.structureVariableOutputData = 0;
- case 5:
- err = (object->*func)( input[0], input[1], input[2],
- input[3], input[4],
- output );
- break;
- case 4:
- err = (object->*func)( input[0], input[1], input[2],
- input[3],
- output, (void *)outputCount );
- break;
- case 3:
- err = (object->*func)( input[0], input[1], input[2],
- output, (void *)outputCount, 0 );
- break;
- case 2:
- err = (object->*func)( input[0], input[1],
- output, (void *)outputCount, 0, 0 );
- break;
- case 1:
- err = (object->*func)( input[0],
- output, (void *)outputCount, 0, 0, 0 );
- break;
- case 0:
- err = (object->*func)( output, (void *)outputCount, 0, 0, 0, 0 );
- break;
+ args.scalarInput = scalar_input;
+ args.scalarInputCount = scalar_inputCnt;
+ args.structureInput = inband_input;
+ args.structureInputSize = inband_inputCnt;
- default:
- IOLog("%s: Bad method table\n", client->getName());
- }
- } while( false);
+ if (ool_input)
+ inputMD = IOMemoryDescriptor::withAddressRange(ool_input, ool_input_size,
+ kIODirectionOut, current_task());
- } else
- err = kIOReturnUnsupported;
+ args.structureInputDescriptor = inputMD;
- return( err);
+ args.scalarOutput = scalar_output;
+ args.scalarOutputCount = *scalar_outputCnt;
+ bzero(&scalar_output[0], *scalar_outputCnt * sizeof(scalar_output[0]));
+ args.structureOutput = inband_output;
+ args.structureOutputSize = *inband_outputCnt;
+
+ if (ool_output && ool_output_size)
+ {
+ outputMD = IOMemoryDescriptor::withAddressRange(ool_output, *ool_output_size,
+ kIODirectionIn, current_task());
+ }
+
+ args.structureOutputDescriptor = outputMD;
+ args.structureOutputDescriptorSize = ool_output_size ? *ool_output_size : 0;
+
+ IOStatisticsClientCall();
+ ret = client->externalMethod( selector, &args );
+
+ *scalar_outputCnt = args.scalarOutputCount;
+ *inband_outputCnt = args.structureOutputSize;
+ *ool_output_size = args.structureOutputDescriptorSize;
+
+ if (inputMD)
+ inputMD->release();
+ if (outputMD)
+ outputMD->release();
+
+ return (ret);
}
-/* Routine io_connect_method_scalarI_structureI */
-kern_return_t is_io_connect_method_scalarI_structureI(
- io_connect_t connect,
- UInt32 index,
- void * input[],
- IOByteCount inputCount,
- UInt8 * inputStruct,
- IOByteCount inputStructCount )
+/* Routine io_async_user_client_method */
+kern_return_t is_io_connect_async_method
+(
+ io_connect_t connection,
+ mach_port_t wake_port,
+ io_async_ref64_t reference,
+ mach_msg_type_number_t referenceCnt,
+ uint32_t selector,
+ io_scalar_inband64_t scalar_input,
+ mach_msg_type_number_t scalar_inputCnt,
+ io_struct_inband_t inband_input,
+ mach_msg_type_number_t inband_inputCnt,
+ mach_vm_address_t ool_input,
+ mach_vm_size_t ool_input_size,
+ io_struct_inband_t inband_output,
+ mach_msg_type_number_t *inband_outputCnt,
+ io_scalar_inband64_t scalar_output,
+ mach_msg_type_number_t *scalar_outputCnt,
+ mach_vm_address_t ool_output,
+ mach_vm_size_t * ool_output_size
+)
+{
+ CHECK( IOUserClient, connection, client );
+
+ IOExternalMethodArguments args;
+ IOReturn ret;
+ IOMemoryDescriptor * inputMD = 0;
+ IOMemoryDescriptor * outputMD = 0;
+
+ bzero(&args.__reserved[0], sizeof(args.__reserved));
+ args.version = kIOExternalMethodArgumentsCurrentVersion;
+
+ reference[0] = (io_user_reference_t) wake_port;
+ if (vm_map_is_64bit(get_task_map(current_task())))
+ reference[0] |= kIOUCAsync64Flag;
+
+ args.selector = selector;
+
+ args.asyncWakePort = wake_port;
+ args.asyncReference = reference;
+ args.asyncReferenceCount = referenceCnt;
+
+ args.scalarInput = scalar_input;
+ args.scalarInputCount = scalar_inputCnt;
+ args.structureInput = inband_input;
+ args.structureInputSize = inband_inputCnt;
+
+ if (ool_input)
+ inputMD = IOMemoryDescriptor::withAddressRange(ool_input, ool_input_size,
+ kIODirectionOut, current_task());
+
+ args.structureInputDescriptor = inputMD;
+
+ args.scalarOutput = scalar_output;
+ args.scalarOutputCount = *scalar_outputCnt;
+ bzero(&scalar_output[0], *scalar_outputCnt * sizeof(scalar_output[0]));
+ args.structureOutput = inband_output;
+ args.structureOutputSize = *inband_outputCnt;
+
+ if (ool_output)
+ {
+ outputMD = IOMemoryDescriptor::withAddressRange(ool_output, *ool_output_size,
+ kIODirectionIn, current_task());
+ }
+
+ args.structureOutputDescriptor = outputMD;
+ args.structureOutputDescriptorSize = *ool_output_size;
+
+ IOStatisticsClientCall();
+ ret = client->externalMethod( selector, &args );
+
+ *inband_outputCnt = args.structureOutputSize;
+ *ool_output_size = args.structureOutputDescriptorSize;
+
+ if (inputMD)
+ inputMD->release();
+ if (outputMD)
+ outputMD->release();
+
+ return (ret);
+}
+
+/* Routine io_connect_method_scalarI_scalarO */
+kern_return_t is_io_connect_method_scalarI_scalarO(
+ io_object_t connect,
+ uint32_t index,
+ io_scalar_inband_t input,
+ mach_msg_type_number_t inputCount,
+ io_scalar_inband_t output,
+ mach_msg_type_number_t * outputCount )
+{
+ IOReturn err;
+ uint32_t i;
+ io_scalar_inband64_t _input;
+ io_scalar_inband64_t _output;
+
+ mach_msg_type_number_t struct_outputCnt = 0;
+ mach_vm_size_t ool_output_size = 0;
+
+ bzero(&_output[0], sizeof(_output));
+ for (i = 0; i < inputCount; i++)
+ _input[i] = SCALAR64(input[i]);
+
+ err = is_io_connect_method(connect, index,
+ _input, inputCount,
+ NULL, 0,
+ 0, 0,
+ NULL, &struct_outputCnt,
+ _output, outputCount,
+ 0, &ool_output_size);
+
+ for (i = 0; i < *outputCount; i++)
+ output[i] = SCALAR32(_output[i]);
+
+ return (err);
+}
+
+kern_return_t shim_io_connect_method_scalarI_scalarO(
+ IOExternalMethod * method,
+ IOService * object,
+ const io_user_scalar_t * input,
+ mach_msg_type_number_t inputCount,
+ io_user_scalar_t * output,
+ mach_msg_type_number_t * outputCount )
{
- IOReturn err;
- IOExternalMethod * method;
- IOService * object;
IOMethod func;
+ io_scalar_inband_t _output;
+ IOReturn err;
+ err = kIOReturnBadArgument;
- CHECK( IOUserClient, connect, client);
+ bzero(&_output[0], sizeof(_output));
+ do {
- if( (method = client->getTargetAndMethodForIndex(&object, index)) ) {
- do {
- err = kIOReturnBadArgument;
- if( kIOUCScalarIStructI != (method->flags & kIOUCTypeMask))
- continue;
- if( (0xffffffff != method->count0)
- && (inputCount != method->count0))
+ if( inputCount != method->count0)
+ {
+ IOLog("%s: IOUserClient inputCount count mismatch\n", object->getName());
continue;
- if( (0xffffffff != method->count1)
- && (inputStructCount != method->count1))
+ }
+ if( *outputCount != method->count1)
+ {
+ IOLog("%s: IOUserClient outputCount count mismatch\n", object->getName());
continue;
+ }
func = method->func;
switch( inputCount) {
+ case 6:
+ err = (object->*func)( ARG32(input[0]), ARG32(input[1]), ARG32(input[2]),
+ ARG32(input[3]), ARG32(input[4]), ARG32(input[5]) );
+ break;
case 5:
- err = (object->*func)( input[0], input[1], input[2],
- input[3], input[4],
- inputStruct );
+ err = (object->*func)( ARG32(input[0]), ARG32(input[1]), ARG32(input[2]),
+ ARG32(input[3]), ARG32(input[4]),
+ &_output[0] );
break;
case 4:
- err = (object->*func)( input[0], input[1], input[2],
- input[3],
- inputStruct, (void *)inputStructCount );
+ err = (object->*func)( ARG32(input[0]), ARG32(input[1]), ARG32(input[2]),
+ ARG32(input[3]),
+ &_output[0], &_output[1] );
break;
case 3:
- err = (object->*func)( input[0], input[1], input[2],
- inputStruct, (void *)inputStructCount,
- 0 );
+ err = (object->*func)( ARG32(input[0]), ARG32(input[1]), ARG32(input[2]),
+ &_output[0], &_output[1], &_output[2] );
break;
case 2:
- err = (object->*func)( input[0], input[1],
- inputStruct, (void *)inputStructCount,
- 0, 0 );
+ err = (object->*func)( ARG32(input[0]), ARG32(input[1]),
+ &_output[0], &_output[1], &_output[2],
+ &_output[3] );
break;
case 1:
- err = (object->*func)( input[0],
- inputStruct, (void *)inputStructCount,
- 0, 0, 0 );
+ err = (object->*func)( ARG32(input[0]),
+ &_output[0], &_output[1], &_output[2],
+ &_output[3], &_output[4] );
break;
case 0:
- err = (object->*func)( inputStruct, (void *)inputStructCount,
- 0, 0, 0, 0 );
+ err = (object->*func)( &_output[0], &_output[1], &_output[2],
+ &_output[3], &_output[4], &_output[5] );
break;
default:
- IOLog("%s: Bad method table\n", client->getName());
+ IOLog("%s: Bad method table\n", object->getName());
}
- } while( false);
+ }
+ while( false);
- } else
- err = kIOReturnUnsupported;
+ uint32_t i;
+ for (i = 0; i < *outputCount; i++)
+ output[i] = SCALAR32(_output[i]);
return( err);
}
-/* Routine io_connect_method_structureI_structureO */
-kern_return_t is_io_connect_method_structureI_structureO(
+/* Routine io_async_method_scalarI_scalarO */
+kern_return_t is_io_async_method_scalarI_scalarO(
+ io_object_t connect,
+ mach_port_t wake_port,
+ io_async_ref_t reference,
+ mach_msg_type_number_t referenceCnt,
+ uint32_t index,
+ io_scalar_inband_t input,
+ mach_msg_type_number_t inputCount,
+ io_scalar_inband_t output,
+ mach_msg_type_number_t * outputCount )
+{
+ IOReturn err;
+ uint32_t i;
+ io_scalar_inband64_t _input;
+ io_scalar_inband64_t _output;
+ io_async_ref64_t _reference;
+
+ bzero(&_output[0], sizeof(_output));
+ for (i = 0; i < referenceCnt; i++)
+ _reference[i] = REF64(reference[i]);
+
+ mach_msg_type_number_t struct_outputCnt = 0;
+ mach_vm_size_t ool_output_size = 0;
+
+ for (i = 0; i < inputCount; i++)
+ _input[i] = SCALAR64(input[i]);
+
+ err = is_io_connect_async_method(connect,
+ wake_port, _reference, referenceCnt,
+ index,
+ _input, inputCount,
+ NULL, 0,
+ 0, 0,
+ NULL, &struct_outputCnt,
+ _output, outputCount,
+ 0, &ool_output_size);
+
+ for (i = 0; i < *outputCount; i++)
+ output[i] = SCALAR32(_output[i]);
+
+ return (err);
+}
+/* Routine io_async_method_scalarI_structureO */
+kern_return_t is_io_async_method_scalarI_structureO(
io_object_t connect,
- UInt32 index,
- UInt8 * input,
- IOByteCount inputCount,
- UInt8 * output,
- IOByteCount * outputCount )
+ mach_port_t wake_port,
+ io_async_ref_t reference,
+ mach_msg_type_number_t referenceCnt,
+ uint32_t index,
+ io_scalar_inband_t input,
+ mach_msg_type_number_t inputCount,
+ io_struct_inband_t output,
+ mach_msg_type_number_t * outputCount )
{
- IOReturn err;
- IOExternalMethod * method;
- IOService * object;
- IOMethod func;
+ uint32_t i;
+ io_scalar_inband64_t _input;
+ io_async_ref64_t _reference;
- CHECK( IOUserClient, connect, client);
+ for (i = 0; i < referenceCnt; i++)
+ _reference[i] = REF64(reference[i]);
- if( (method = client->getTargetAndMethodForIndex(&object, index)) ) {
- do {
- err = kIOReturnBadArgument;
- if( kIOUCStructIStructO != (method->flags & kIOUCTypeMask))
- continue;
- if( (0xffffffff != method->count0)
- && (inputCount != method->count0))
- continue;
- if( (0xffffffff != method->count1)
- && (*outputCount != method->count1))
- continue;
+ mach_msg_type_number_t scalar_outputCnt = 0;
+ mach_vm_size_t ool_output_size = 0;
- func = method->func;
+ for (i = 0; i < inputCount; i++)
+ _input[i] = SCALAR64(input[i]);
+
+ return (is_io_connect_async_method(connect,
+ wake_port, _reference, referenceCnt,
+ index,
+ _input, inputCount,
+ NULL, 0,
+ 0, 0,
+ output, outputCount,
+ NULL, &scalar_outputCnt,
+ 0, &ool_output_size));
+}
- if( method->count1) {
- if( method->count0) {
- err = (object->*func)( input, output,
- (void *)inputCount, outputCount, 0, 0 );
- } else {
- err = (object->*func)( output, outputCount, 0, 0, 0, 0 );
- }
- } else {
- err = (object->*func)( input, (void *)inputCount, 0, 0, 0, 0 );
- }
+/* Routine io_async_method_scalarI_structureI */
+kern_return_t is_io_async_method_scalarI_structureI(
+ io_connect_t connect,
+ mach_port_t wake_port,
+ io_async_ref_t reference,
+ mach_msg_type_number_t referenceCnt,
+ uint32_t index,
+ io_scalar_inband_t input,
+ mach_msg_type_number_t inputCount,
+ io_struct_inband_t inputStruct,
+ mach_msg_type_number_t inputStructCount )
+{
+ uint32_t i;
+ io_scalar_inband64_t _input;
+ io_async_ref64_t _reference;
- } while( false);
+ for (i = 0; i < referenceCnt; i++)
+ _reference[i] = REF64(reference[i]);
- } else
- err = kIOReturnUnsupported;
+ mach_msg_type_number_t scalar_outputCnt = 0;
+ mach_msg_type_number_t inband_outputCnt = 0;
+ mach_vm_size_t ool_output_size = 0;
- return( err);
+ for (i = 0; i < inputCount; i++)
+ _input[i] = SCALAR64(input[i]);
+
+ return (is_io_connect_async_method(connect,
+ wake_port, _reference, referenceCnt,
+ index,
+ _input, inputCount,
+ inputStruct, inputStructCount,
+ 0, 0,
+ NULL, &inband_outputCnt,
+ NULL, &scalar_outputCnt,
+ 0, &ool_output_size));
}
-kern_return_t is_io_async_method_scalarI_scalarO(
- io_object_t connect,
- mach_port_t wakePort,
- io_async_ref_t reference,
- mach_msg_type_number_t referenceCnt,
- UInt32 index,
- void * input[],
- IOByteCount inputCount,
- void * output[],
- IOByteCount * outputCount )
+/* Routine io_async_method_structureI_structureO */
+kern_return_t is_io_async_method_structureI_structureO(
+ io_object_t connect,
+ mach_port_t wake_port,
+ io_async_ref_t reference,
+ mach_msg_type_number_t referenceCnt,
+ uint32_t index,
+ io_struct_inband_t input,
+ mach_msg_type_number_t inputCount,
+ io_struct_inband_t output,
+ mach_msg_type_number_t * outputCount )
+{
+ uint32_t i;
+ mach_msg_type_number_t scalar_outputCnt = 0;
+ mach_vm_size_t ool_output_size = 0;
+ io_async_ref64_t _reference;
+
+ for (i = 0; i < referenceCnt; i++)
+ _reference[i] = REF64(reference[i]);
+
+ return (is_io_connect_async_method(connect,
+ wake_port, _reference, referenceCnt,
+ index,
+ NULL, 0,
+ input, inputCount,
+ 0, 0,
+ output, outputCount,
+ NULL, &scalar_outputCnt,
+ 0, &ool_output_size));
+}
+
+
+kern_return_t shim_io_async_method_scalarI_scalarO(
+ IOExternalAsyncMethod * method,
+ IOService * object,
+ mach_port_t asyncWakePort,
+ io_user_reference_t * asyncReference,
+ uint32_t asyncReferenceCount,
+ const io_user_scalar_t * input,
+ mach_msg_type_number_t inputCount,
+ io_user_scalar_t * output,
+ mach_msg_type_number_t * outputCount )
{
- IOReturn err;
- IOExternalAsyncMethod *method;
- IOService * object;
IOAsyncMethod func;
+ uint32_t i;
+ io_scalar_inband_t _output;
+ IOReturn err;
+ io_async_ref_t reference;
- CHECK( IOUserClient, connect, client);
- if( (method = client->getAsyncTargetAndMethodForIndex(&object, index)) ) {
- do {
- err = kIOReturnBadArgument;
- if( kIOUCScalarIScalarO != (method->flags & kIOUCTypeMask))
- continue;
- if( inputCount != method->count0)
- continue;
- if( *outputCount != method->count1)
- continue;
+ bzero(&_output[0], sizeof(_output));
+ for (i = 0; i < asyncReferenceCount; i++)
+ reference[i] = REF32(asyncReference[i]);
- reference[0] = (natural_t) wakePort;
- func = method->func;
+ err = kIOReturnBadArgument;
+
+ do {
+
+ if( inputCount != method->count0)
+ {
+ IOLog("%s: IOUserClient inputCount count mismatch\n", object->getName());
+ continue;
+ }
+ if( *outputCount != method->count1)
+ {
+ IOLog("%s: IOUserClient outputCount count mismatch\n", object->getName());
+ continue;
+ }
+
+ func = method->func;
switch( inputCount) {
case 6:
err = (object->*func)( reference,
- input[0], input[1], input[2],
- input[3], input[4], input[5] );
+ ARG32(input[0]), ARG32(input[1]), ARG32(input[2]),
+ ARG32(input[3]), ARG32(input[4]), ARG32(input[5]) );
break;
case 5:
err = (object->*func)( reference,
- input[0], input[1], input[2],
- input[3], input[4],
- &output[0] );
+ ARG32(input[0]), ARG32(input[1]), ARG32(input[2]),
+ ARG32(input[3]), ARG32(input[4]),
+ &_output[0] );
break;
case 4:
err = (object->*func)( reference,
- input[0], input[1], input[2],
- input[3],
- &output[0], &output[1] );
+ ARG32(input[0]), ARG32(input[1]), ARG32(input[2]),
+ ARG32(input[3]),
+ &_output[0], &_output[1] );
break;
case 3:
err = (object->*func)( reference,
- input[0], input[1], input[2],
- &output[0], &output[1], &output[2] );
+ ARG32(input[0]), ARG32(input[1]), ARG32(input[2]),
+ &_output[0], &_output[1], &_output[2] );
break;
case 2:
err = (object->*func)( reference,
- input[0], input[1],
- &output[0], &output[1], &output[2],
- &output[3] );
+ ARG32(input[0]), ARG32(input[1]),
+ &_output[0], &_output[1], &_output[2],
+ &_output[3] );
break;
case 1:
err = (object->*func)( reference,
- input[0],
- &output[0], &output[1], &output[2],
- &output[3], &output[4] );
+ ARG32(input[0]),
+ &_output[0], &_output[1], &_output[2],
+ &_output[3], &_output[4] );
break;
case 0:
err = (object->*func)( reference,
- &output[0], &output[1], &output[2],
- &output[3], &output[4], &output[5] );
+ &_output[0], &_output[1], &_output[2],
+ &_output[3], &_output[4], &_output[5] );
break;
default:
- IOLog("%s: Bad method table\n", client->getName());
+ IOLog("%s: Bad method table\n", object->getName());
}
- } while( false);
+ }
+ while( false);
- } else
- err = kIOReturnUnsupported;
+ for (i = 0; i < *outputCount; i++)
+ output[i] = SCALAR32(_output[i]);
return( err);
}
-kern_return_t is_io_async_method_scalarI_structureO(
- io_object_t connect,
- mach_port_t wakePort,
- io_async_ref_t reference,
- mach_msg_type_number_t referenceCnt,
- UInt32 index,
- void * input[],
- IOByteCount inputCount,
- void * output,
+
+/* Routine io_connect_method_scalarI_structureO */
+kern_return_t is_io_connect_method_scalarI_structureO(
+ io_object_t connect,
+ uint32_t index,
+ io_scalar_inband_t input,
+ mach_msg_type_number_t inputCount,
+ io_struct_inband_t output,
+ mach_msg_type_number_t * outputCount )
+{
+ uint32_t i;
+ io_scalar_inband64_t _input;
+
+ mach_msg_type_number_t scalar_outputCnt = 0;
+ mach_vm_size_t ool_output_size = 0;
+
+ for (i = 0; i < inputCount; i++)
+ _input[i] = SCALAR64(input[i]);
+
+ return (is_io_connect_method(connect, index,
+ _input, inputCount,
+ NULL, 0,
+ 0, 0,
+ output, outputCount,
+ NULL, &scalar_outputCnt,
+ 0, &ool_output_size));
+}
+
+kern_return_t shim_io_connect_method_scalarI_structureO(
+
+ IOExternalMethod * method,
+ IOService * object,
+ const io_user_scalar_t * input,
+ mach_msg_type_number_t inputCount,
+ io_struct_inband_t output,
IOByteCount * outputCount )
{
+ IOMethod func;
IOReturn err;
- IOExternalAsyncMethod *method;
- IOService * object;
+
+ err = kIOReturnBadArgument;
+
+ do {
+ if( inputCount != method->count0)
+ {
+ IOLog("%s: IOUserClient inputCount count mismatch\n", object->getName());
+ continue;
+ }
+ if( (kIOUCVariableStructureSize != method->count1)
+ && (*outputCount != method->count1))
+ {
+ IOLog("%s: IOUserClient outputCount count mismatch\n", object->getName());
+ continue;
+ }
+
+ func = method->func;
+
+ switch( inputCount) {
+
+ case 5:
+ err = (object->*func)( ARG32(input[0]), ARG32(input[1]), ARG32(input[2]),
+ ARG32(input[3]), ARG32(input[4]),
+ output );
+ break;
+ case 4:
+ err = (object->*func)( ARG32(input[0]), ARG32(input[1]), ARG32(input[2]),
+ ARG32(input[3]),
+ output, (void *)outputCount );
+ break;
+ case 3:
+ err = (object->*func)( ARG32(input[0]), ARG32(input[1]), ARG32(input[2]),
+ output, (void *)outputCount, 0 );
+ break;
+ case 2:
+ err = (object->*func)( ARG32(input[0]), ARG32(input[1]),
+ output, (void *)outputCount, 0, 0 );
+ break;
+ case 1:
+ err = (object->*func)( ARG32(input[0]),
+ output, (void *)outputCount, 0, 0, 0 );
+ break;
+ case 0:
+ err = (object->*func)( output, (void *)outputCount, 0, 0, 0, 0 );
+ break;
+
+ default:
+ IOLog("%s: Bad method table\n", object->getName());
+ }
+ }
+ while( false);
+
+ return( err);
+}
+
+
+kern_return_t shim_io_async_method_scalarI_structureO(
+ IOExternalAsyncMethod * method,
+ IOService * object,
+ mach_port_t asyncWakePort,
+ io_user_reference_t * asyncReference,
+ uint32_t asyncReferenceCount,
+ const io_user_scalar_t * input,
+ mach_msg_type_number_t inputCount,
+ io_struct_inband_t output,
+ mach_msg_type_number_t * outputCount )
+{
IOAsyncMethod func;
+ uint32_t i;
+ IOReturn err;
+ io_async_ref_t reference;
- CHECK( IOUserClient, connect, client);
+ for (i = 0; i < asyncReferenceCount; i++)
+ reference[i] = REF32(asyncReference[i]);
- if( (method = client->getAsyncTargetAndMethodForIndex(&object, index)) ) {
- do {
- err = kIOReturnBadArgument;
- if( kIOUCScalarIStructO != (method->flags & kIOUCTypeMask))
- continue;
- if( inputCount != method->count0)
- continue;
- if( (0xffffffff != method->count1)
- && (*outputCount != method->count1))
- continue;
+ err = kIOReturnBadArgument;
+ do {
+ if( inputCount != method->count0)
+ {
+ IOLog("%s: IOUserClient inputCount count mismatch\n", object->getName());
+ continue;
+ }
+ if( (kIOUCVariableStructureSize != method->count1)
+ && (*outputCount != method->count1))
+ {
+ IOLog("%s: IOUserClient outputCount count mismatch\n", object->getName());
+ continue;
+ }
- reference[0] = (natural_t) wakePort;
- func = method->func;
+ func = method->func;
switch( inputCount) {
case 5:
err = (object->*func)( reference,
- input[0], input[1], input[2],
- input[3], input[4],
+ ARG32(input[0]), ARG32(input[1]), ARG32(input[2]),
+ ARG32(input[3]), ARG32(input[4]),
output );
break;
case 4:
err = (object->*func)( reference,
- input[0], input[1], input[2],
- input[3],
+ ARG32(input[0]), ARG32(input[1]), ARG32(input[2]),
+ ARG32(input[3]),
output, (void *)outputCount );
break;
case 3:
err = (object->*func)( reference,
- input[0], input[1], input[2],
+ ARG32(input[0]), ARG32(input[1]), ARG32(input[2]),
output, (void *)outputCount, 0 );
break;
case 2:
err = (object->*func)( reference,
- input[0], input[1],
+ ARG32(input[0]), ARG32(input[1]),
output, (void *)outputCount, 0, 0 );
break;
case 1:
err = (object->*func)( reference,
- input[0],
+ ARG32(input[0]),
output, (void *)outputCount, 0, 0, 0 );
break;
case 0:
break;
default:
- IOLog("%s: Bad method table\n", client->getName());
+ IOLog("%s: Bad method table\n", object->getName());
}
- } while( false);
+ }
+ while( false);
- } else
- err = kIOReturnUnsupported;
+ return( err);
+}
+
+/* Routine io_connect_method_scalarI_structureI */
+kern_return_t is_io_connect_method_scalarI_structureI(
+ io_connect_t connect,
+ uint32_t index,
+ io_scalar_inband_t input,
+ mach_msg_type_number_t inputCount,
+ io_struct_inband_t inputStruct,
+ mach_msg_type_number_t inputStructCount )
+{
+ uint32_t i;
+ io_scalar_inband64_t _input;
+
+ mach_msg_type_number_t scalar_outputCnt = 0;
+ mach_msg_type_number_t inband_outputCnt = 0;
+ mach_vm_size_t ool_output_size = 0;
+
+ for (i = 0; i < inputCount; i++)
+ _input[i] = SCALAR64(input[i]);
+
+ return (is_io_connect_method(connect, index,
+ _input, inputCount,
+ inputStruct, inputStructCount,
+ 0, 0,
+ NULL, &inband_outputCnt,
+ NULL, &scalar_outputCnt,
+ 0, &ool_output_size));
+}
+
+kern_return_t shim_io_connect_method_scalarI_structureI(
+ IOExternalMethod * method,
+ IOService * object,
+ const io_user_scalar_t * input,
+ mach_msg_type_number_t inputCount,
+ io_struct_inband_t inputStruct,
+ mach_msg_type_number_t inputStructCount )
+{
+ IOMethod func;
+ IOReturn err = kIOReturnBadArgument;
+
+ do
+ {
+ if( (kIOUCVariableStructureSize != method->count0)
+ && (inputCount != method->count0))
+ {
+ IOLog("%s: IOUserClient inputCount count mismatch\n", object->getName());
+ continue;
+ }
+ if( (kIOUCVariableStructureSize != method->count1)
+ && (inputStructCount != method->count1))
+ {
+ IOLog("%s: IOUserClient outputCount count mismatch\n", object->getName());
+ continue;
+ }
+
+ func = method->func;
+
+ switch( inputCount) {
+
+ case 5:
+ err = (object->*func)( ARG32(input[0]), ARG32(input[1]), ARG32(input[2]),
+ ARG32(input[3]), ARG32(input[4]),
+ inputStruct );
+ break;
+ case 4:
+ err = (object->*func)( ARG32(input[0]), ARG32(input[1]), (void *) input[2],
+ ARG32(input[3]),
+ inputStruct, (void *)(uintptr_t)inputStructCount );
+ break;
+ case 3:
+ err = (object->*func)( ARG32(input[0]), ARG32(input[1]), ARG32(input[2]),
+ inputStruct, (void *)(uintptr_t)inputStructCount,
+ 0 );
+ break;
+ case 2:
+ err = (object->*func)( ARG32(input[0]), ARG32(input[1]),
+ inputStruct, (void *)(uintptr_t)inputStructCount,
+ 0, 0 );
+ break;
+ case 1:
+ err = (object->*func)( ARG32(input[0]),
+ inputStruct, (void *)(uintptr_t)inputStructCount,
+ 0, 0, 0 );
+ break;
+ case 0:
+ err = (object->*func)( inputStruct, (void *)(uintptr_t)inputStructCount,
+ 0, 0, 0, 0 );
+ break;
+
+ default:
+ IOLog("%s: Bad method table\n", object->getName());
+ }
+ }
+ while (false);
return( err);
}
-kern_return_t is_io_async_method_scalarI_structureI(
- io_connect_t connect,
- mach_port_t wakePort,
- io_async_ref_t reference,
- mach_msg_type_number_t referenceCnt,
- UInt32 index,
- void * input[],
- IOByteCount inputCount,
- UInt8 * inputStruct,
- IOByteCount inputStructCount )
+kern_return_t shim_io_async_method_scalarI_structureI(
+ IOExternalAsyncMethod * method,
+ IOService * object,
+ mach_port_t asyncWakePort,
+ io_user_reference_t * asyncReference,
+ uint32_t asyncReferenceCount,
+ const io_user_scalar_t * input,
+ mach_msg_type_number_t inputCount,
+ io_struct_inband_t inputStruct,
+ mach_msg_type_number_t inputStructCount )
{
- IOReturn err;
- IOExternalAsyncMethod *method;
- IOService * object;
IOAsyncMethod func;
+ uint32_t i;
+ IOReturn err = kIOReturnBadArgument;
+ io_async_ref_t reference;
- CHECK( IOUserClient, connect, client);
+ for (i = 0; i < asyncReferenceCount; i++)
+ reference[i] = REF32(asyncReference[i]);
- if( (method = client->getAsyncTargetAndMethodForIndex(&object, index)) ) {
- do {
- err = kIOReturnBadArgument;
- if( kIOUCScalarIStructI != (method->flags & kIOUCTypeMask))
- continue;
- if( (0xffffffff != method->count0)
- && (inputCount != method->count0))
- continue;
- if( (0xffffffff != method->count1)
- && (inputStructCount != method->count1))
- continue;
+ do
+ {
+ if( (kIOUCVariableStructureSize != method->count0)
+ && (inputCount != method->count0))
+ {
+ IOLog("%s: IOUserClient inputCount count mismatch\n", object->getName());
+ continue;
+ }
+ if( (kIOUCVariableStructureSize != method->count1)
+ && (inputStructCount != method->count1))
+ {
+ IOLog("%s: IOUserClient outputCount count mismatch\n", object->getName());
+ continue;
+ }
- reference[0] = (natural_t) wakePort;
func = method->func;
switch( inputCount) {
case 5:
err = (object->*func)( reference,
- input[0], input[1], input[2],
- input[3], input[4],
+ ARG32(input[0]), ARG32(input[1]), ARG32(input[2]),
+ ARG32(input[3]), ARG32(input[4]),
inputStruct );
break;
case 4:
err = (object->*func)( reference,
- input[0], input[1], input[2],
- input[3],
- inputStruct, (void *)inputStructCount );
+ ARG32(input[0]), ARG32(input[1]), ARG32(input[2]),
+ ARG32(input[3]),
+ inputStruct, (void *)(uintptr_t)inputStructCount );
break;
case 3:
err = (object->*func)( reference,
- input[0], input[1], input[2],
- inputStruct, (void *)inputStructCount,
+ ARG32(input[0]), ARG32(input[1]), ARG32(input[2]),
+ inputStruct, (void *)(uintptr_t)inputStructCount,
0 );
break;
case 2:
err = (object->*func)( reference,
- input[0], input[1],
- inputStruct, (void *)inputStructCount,
+ ARG32(input[0]), ARG32(input[1]),
+ inputStruct, (void *)(uintptr_t)inputStructCount,
0, 0 );
break;
case 1:
err = (object->*func)( reference,
- input[0],
- inputStruct, (void *)inputStructCount,
+ ARG32(input[0]),
+ inputStruct, (void *)(uintptr_t)inputStructCount,
0, 0, 0 );
break;
case 0:
err = (object->*func)( reference,
- inputStruct, (void *)inputStructCount,
+ inputStruct, (void *)(uintptr_t)inputStructCount,
0, 0, 0, 0 );
break;
default:
- IOLog("%s: Bad method table\n", client->getName());
+ IOLog("%s: Bad method table\n", object->getName());
}
- } while( false);
-
- } else
- err = kIOReturnUnsupported;
+ }
+ while (false);
return( err);
}
-kern_return_t is_io_async_method_structureI_structureO(
- io_object_t connect,
- mach_port_t wakePort,
- io_async_ref_t reference,
- mach_msg_type_number_t referenceCnt,
- UInt32 index,
- UInt8 * input,
- IOByteCount inputCount,
- UInt8 * output,
+/* Routine io_connect_method_structureI_structureO */
+kern_return_t is_io_connect_method_structureI_structureO(
+ io_object_t connect,
+ uint32_t index,
+ io_struct_inband_t input,
+ mach_msg_type_number_t inputCount,
+ io_struct_inband_t output,
+ mach_msg_type_number_t * outputCount )
+{
+ mach_msg_type_number_t scalar_outputCnt = 0;
+ mach_vm_size_t ool_output_size = 0;
+
+ return (is_io_connect_method(connect, index,
+ NULL, 0,
+ input, inputCount,
+ 0, 0,
+ output, outputCount,
+ NULL, &scalar_outputCnt,
+ 0, &ool_output_size));
+}
+
+kern_return_t shim_io_connect_method_structureI_structureO(
+ IOExternalMethod * method,
+ IOService * object,
+ io_struct_inband_t input,
+ mach_msg_type_number_t inputCount,
+ io_struct_inband_t output,
IOByteCount * outputCount )
{
- IOReturn err;
- IOExternalAsyncMethod *method;
- IOService * object;
+ IOMethod func;
+ IOReturn err = kIOReturnBadArgument;
+
+ do
+ {
+ if( (kIOUCVariableStructureSize != method->count0)
+ && (inputCount != method->count0))
+ {
+ IOLog("%s: IOUserClient inputCount count mismatch\n", object->getName());
+ continue;
+ }
+ if( (kIOUCVariableStructureSize != method->count1)
+ && (*outputCount != method->count1))
+ {
+ IOLog("%s: IOUserClient outputCount count mismatch\n", object->getName());
+ continue;
+ }
+
+ func = method->func;
+
+ if( method->count1) {
+ if( method->count0) {
+ err = (object->*func)( input, output,
+ (void *)(uintptr_t)inputCount, outputCount, 0, 0 );
+ } else {
+ err = (object->*func)( output, outputCount, 0, 0, 0, 0 );
+ }
+ } else {
+ err = (object->*func)( input, (void *)(uintptr_t)inputCount, 0, 0, 0, 0 );
+ }
+ }
+ while( false);
+
+
+ return( err);
+}
+
+kern_return_t shim_io_async_method_structureI_structureO(
+ IOExternalAsyncMethod * method,
+ IOService * object,
+ mach_port_t asyncWakePort,
+ io_user_reference_t * asyncReference,
+ uint32_t asyncReferenceCount,
+ io_struct_inband_t input,
+ mach_msg_type_number_t inputCount,
+ io_struct_inband_t output,
+ mach_msg_type_number_t * outputCount )
+{
IOAsyncMethod func;
+ uint32_t i;
+ IOReturn err;
+ io_async_ref_t reference;
- CHECK( IOUserClient, connect, client);
+ for (i = 0; i < asyncReferenceCount; i++)
+ reference[i] = REF32(asyncReference[i]);
- if( (method = client->getAsyncTargetAndMethodForIndex(&object, index)) ) {
- do {
- err = kIOReturnBadArgument;
- if( kIOUCStructIStructO != (method->flags & kIOUCTypeMask))
- continue;
- if( (0xffffffff != method->count0)
- && (inputCount != method->count0))
- continue;
- if( (0xffffffff != method->count1)
- && (*outputCount != method->count1))
- continue;
+ err = kIOReturnBadArgument;
+ do
+ {
+ if( (kIOUCVariableStructureSize != method->count0)
+ && (inputCount != method->count0))
+ {
+ IOLog("%s: IOUserClient inputCount count mismatch\n", object->getName());
+ continue;
+ }
+ if( (kIOUCVariableStructureSize != method->count1)
+ && (*outputCount != method->count1))
+ {
+ IOLog("%s: IOUserClient outputCount count mismatch\n", object->getName());
+ continue;
+ }
- reference[0] = (natural_t) wakePort;
func = method->func;
if( method->count1) {
if( method->count0) {
err = (object->*func)( reference,
input, output,
- (void *)inputCount, outputCount, 0, 0 );
+ (void *)(uintptr_t)inputCount, outputCount, 0, 0 );
} else {
err = (object->*func)( reference,
output, outputCount, 0, 0, 0, 0 );
}
} else {
err = (object->*func)( reference,
- input, (void *)inputCount, 0, 0, 0, 0 );
+ input, (void *)(uintptr_t)inputCount, 0, 0, 0, 0 );
}
-
- } while( false);
-
- } else
- err = kIOReturnUnsupported;
-
- return( err);
-}
-/* Routine io_make_matching */
-kern_return_t is_io_make_matching(
- mach_port_t master_port,
- UInt32 type,
- IOOptionBits options,
- UInt8 * input,
- IOByteCount inputCount,
- io_string_t matching )
-{
- OSSerialize * s;
- IOReturn err = kIOReturnSuccess;
- OSDictionary * dict;
-
- if( master_port != master_device_port)
- return( kIOReturnNotPrivileged);
-
- switch( type) {
-
- case kIOServiceMatching:
- dict = IOService::serviceMatching( gIOServiceKey );
- break;
-
- case kIOBSDNameMatching:
- dict = IOBSDNameMatching( (const char *) input );
- break;
-
- case kIOOFPathMatching:
- dict = IOOFPathMatching( (const char *) input,
- matching, sizeof( io_string_t));
- break;
-
- default:
- dict = 0;
}
-
- if( !dict)
- return( kIOReturnUnsupported);
-
- do {
- s = OSSerialize::withCapacity(4096);
- if( !s) {
- err = kIOReturnNoMemory;
- continue;
- }
- s->clearText();
- if( !dict->serialize( s )) {
- err = kIOReturnUnsupported;
- continue;
- }
-
- if( s->getLength() > sizeof( io_string_t)) {
- err = kIOReturnNoMemory;
- continue;
- } else
- strcpy( matching, s->text());
-
- } while( false);
-
- if( s)
- s->release();
- if( dict)
- dict->release();
+ while( false);
return( err);
}
/* Routine io_catalog_send_data */
kern_return_t is_io_catalog_send_data(
mach_port_t master_port,
- int flag,
+ uint32_t flag,
io_buf_ptr_t inData,
mach_msg_type_number_t inDataCount,
- natural_t * result)
+ kern_return_t * result)
{
OSObject * obj = 0;
vm_offset_t data;
if( master_port != master_device_port)
return kIOReturnNotPrivileged;
- // FIXME: This is a hack. Should have own function for removeKernelLinker()
- if(flag != kIOCatalogRemoveKernelLinker && ( !inData || !inDataCount) )
+ if( (flag != kIOCatalogRemoveKernelLinker &&
+ flag != kIOCatalogKextdActive &&
+ flag != kIOCatalogKextdFinishedLaunching) &&
+ ( !inData || !inDataCount) )
+ {
return kIOReturnBadArgument;
+ }
if (inData) {
vm_map_offset_t map_data;
+ if( inDataCount > sizeof(io_struct_inband_t) * 1024)
+ return( kIOReturnMessageTooLarge);
+
kr = vm_map_copyout( kernel_map, &map_data, (vm_map_copy_t)inData);
- data = CAST_DOWN(vm_offset_t, map_data);
+ data = CAST_DOWN(vm_offset_t, map_data);
if( kr != KERN_SUCCESS)
return kr;
// must return success after vm_map_copyout() succeeds
if( inDataCount ) {
- obj = (OSObject *)OSUnserializeXML((const char *)data);
+ obj = (OSObject *)OSUnserializeXML((const char *)data, inDataCount);
vm_deallocate( kernel_map, data, inDataCount );
if( !obj) {
*result = kIOReturnNoMemory;
}
switch ( flag ) {
+ case kIOCatalogResetDrivers:
+ case kIOCatalogResetDriversNoMatch: {
+ OSArray * array;
+
+ array = OSDynamicCast(OSArray, obj);
+ if (array) {
+ if ( !gIOCatalogue->resetAndAddDrivers(array,
+ flag == kIOCatalogResetDrivers) ) {
+
+ kr = kIOReturnError;
+ }
+ } else {
+ kr = kIOReturnBadArgument;
+ }
+ }
+ break;
+
case kIOCatalogAddDrivers:
case kIOCatalogAddDriversNoMatch: {
OSArray * array;
}
break;
- case kIOCatalogRemoveKernelLinker: {
- if (gIOCatalogue->removeKernelLinker() != KERN_SUCCESS) {
- kr = kIOReturnError;
- } else {
- kr = kIOReturnSuccess;
+ case kIOCatalogRemoveKernelLinker:
+ kr = KERN_NOT_SUPPORTED;
+ break;
+
+ case kIOCatalogKextdActive:
+#if !NO_KEXTD
+ IOServiceTrace(IOSERVICE_KEXTD_ALIVE, 0, 0, 0, 0);
+ OSKext::setKextdActive();
+
+ /* Dump all nonloaded startup extensions; kextd will now send them
+ * down on request.
+ */
+ OSKext::flushNonloadedKexts( /* flushPrelinkedKexts */ false);
+#endif
+ kr = kIOReturnSuccess;
+ break;
+
+ case kIOCatalogKextdFinishedLaunching: {
+#if !NO_KEXTD
+ static bool clearedBusy = false;
+
+ if (!clearedBusy) {
+ IOService * serviceRoot = IOService::getServiceRoot();
+ if (serviceRoot) {
+ IOServiceTrace(IOSERVICE_KEXTD_READY, 0, 0, 0, 0);
+ serviceRoot->adjustBusy(-1);
+ clearedBusy = true;
+ }
}
+#endif
+ kr = kIOReturnSuccess;
}
break;
/* Routine io_catalog_terminate */
kern_return_t is_io_catalog_terminate(
mach_port_t master_port,
- int flag,
+ uint32_t flag,
io_name_t name )
{
kern_return_t kr;
return( kr );
switch ( flag ) {
+#if !defined(SECURE_KERNEL)
case kIOCatalogServiceTerminate:
OSIterator * iter;
IOService * service;
kr = gIOCatalogue->terminateDriversForModule(name,
flag == kIOCatalogModuleUnload);
break;
+#endif
default:
kr = kIOReturnBadArgument;
/* Routine io_catalog_get_data */
kern_return_t is_io_catalog_get_data(
mach_port_t master_port,
- int flag,
+ uint32_t flag,
io_buf_ptr_t *outData,
mach_msg_type_number_t *outDataCount)
{
if ( !s )
return kIOReturnNoMemory;
- s->clearText();
-
kr = gIOCatalogue->serializeData(flag, s);
if ( kr == kIOReturnSuccess ) {
/* Routine io_catalog_get_gen_count */
kern_return_t is_io_catalog_get_gen_count(
mach_port_t master_port,
- int *genCount)
+ uint32_t *genCount)
{
if( master_port != master_device_port)
return kIOReturnNotPrivileged;
return kIOReturnSuccess;
}
-/* Routine io_catalog_module_loaded */
+/* Routine io_catalog_module_loaded.
+ * Is invoked from IOKitLib's IOCatalogueModuleLoaded(). Doesn't seem to be used.
+ */
kern_return_t is_io_catalog_module_loaded(
mach_port_t master_port,
io_name_t name)
kern_return_t is_io_catalog_reset(
mach_port_t master_port,
- int flag)
+ uint32_t flag)
{
if( master_port != master_device_port)
return kIOReturnNotPrivileged;
return result;
}
-}; /* extern "C" */
+} /* extern "C" */
+
+IOReturn IOUserClient::externalMethod( uint32_t selector, IOExternalMethodArguments * args,
+ IOExternalMethodDispatch * dispatch, OSObject * target, void * reference )
+{
+ IOReturn err;
+ IOService * object;
+ IOByteCount structureOutputSize;
+
+ if (dispatch)
+ {
+ uint32_t count;
+ count = dispatch->checkScalarInputCount;
+ if ((kIOUCVariableStructureSize != count) && (count != args->scalarInputCount))
+ {
+ return (kIOReturnBadArgument);
+ }
+
+ count = dispatch->checkStructureInputSize;
+ if ((kIOUCVariableStructureSize != count)
+ && (count != ((args->structureInputDescriptor)
+ ? args->structureInputDescriptor->getLength() : args->structureInputSize)))
+ {
+ return (kIOReturnBadArgument);
+ }
+
+ count = dispatch->checkScalarOutputCount;
+ if ((kIOUCVariableStructureSize != count) && (count != args->scalarOutputCount))
+ {
+ return (kIOReturnBadArgument);
+ }
+
+ count = dispatch->checkStructureOutputSize;
+ if ((kIOUCVariableStructureSize != count)
+ && (count != ((args->structureOutputDescriptor)
+ ? args->structureOutputDescriptor->getLength() : args->structureOutputSize)))
+ {
+ return (kIOReturnBadArgument);
+ }
+
+ if (dispatch->function)
+ err = (*dispatch->function)(target, reference, args);
+ else
+ err = kIOReturnNoCompletion; /* implementator can dispatch */
+
+ return (err);
+ }
+
+
+ // pre-Leopard API's don't do ool structs
+ if (args->structureInputDescriptor || args->structureOutputDescriptor)
+ {
+ err = kIOReturnIPCError;
+ return (err);
+ }
+
+ structureOutputSize = args->structureOutputSize;
+ if (args->asyncWakePort)
+ {
+ IOExternalAsyncMethod * method;
+
+ if( !(method = getAsyncTargetAndMethodForIndex(&object, selector)) )
+ return (kIOReturnUnsupported);
+
+ if (kIOUCForegroundOnly & method->flags)
+ {
+ if (task_is_gpu_denied(current_task()))
+ return (kIOReturnNotPermitted);
+ }
+
+ switch (method->flags & kIOUCTypeMask)
+ {
+ case kIOUCScalarIStructI:
+ err = shim_io_async_method_scalarI_structureI( method, object,
+ args->asyncWakePort, args->asyncReference, args->asyncReferenceCount,
+ args->scalarInput, args->scalarInputCount,
+ (char *)args->structureInput, args->structureInputSize );
+ break;
+
+ case kIOUCScalarIScalarO:
+ err = shim_io_async_method_scalarI_scalarO( method, object,
+ args->asyncWakePort, args->asyncReference, args->asyncReferenceCount,
+ args->scalarInput, args->scalarInputCount,
+ args->scalarOutput, &args->scalarOutputCount );
+ break;
+
+ case kIOUCScalarIStructO:
+ err = shim_io_async_method_scalarI_structureO( method, object,
+ args->asyncWakePort, args->asyncReference, args->asyncReferenceCount,
+ args->scalarInput, args->scalarInputCount,
+ (char *) args->structureOutput, &args->structureOutputSize );
+ break;
+
+
+ case kIOUCStructIStructO:
+ err = shim_io_async_method_structureI_structureO( method, object,
+ args->asyncWakePort, args->asyncReference, args->asyncReferenceCount,
+ (char *)args->structureInput, args->structureInputSize,
+ (char *) args->structureOutput, &args->structureOutputSize );
+ break;
+
+ default:
+ err = kIOReturnBadArgument;
+ break;
+ }
+ }
+ else
+ {
+ IOExternalMethod * method;
+
+ if( !(method = getTargetAndMethodForIndex(&object, selector)) )
+ return (kIOReturnUnsupported);
+
+ if (kIOUCForegroundOnly & method->flags)
+ {
+ if (task_is_gpu_denied(current_task()))
+ return (kIOReturnNotPermitted);
+ }
+
+ switch (method->flags & kIOUCTypeMask)
+ {
+ case kIOUCScalarIStructI:
+ err = shim_io_connect_method_scalarI_structureI( method, object,
+ args->scalarInput, args->scalarInputCount,
+ (char *) args->structureInput, args->structureInputSize );
+ break;
+
+ case kIOUCScalarIScalarO:
+ err = shim_io_connect_method_scalarI_scalarO( method, object,
+ args->scalarInput, args->scalarInputCount,
+ args->scalarOutput, &args->scalarOutputCount );
+ break;
+
+ case kIOUCScalarIStructO:
+ err = shim_io_connect_method_scalarI_structureO( method, object,
+ args->scalarInput, args->scalarInputCount,
+ (char *) args->structureOutput, &structureOutputSize );
+ break;
+
+
+ case kIOUCStructIStructO:
+ err = shim_io_connect_method_structureI_structureO( method, object,
+ (char *) args->structureInput, args->structureInputSize,
+ (char *) args->structureOutput, &structureOutputSize );
+ break;
+
+ default:
+ err = kIOReturnBadArgument;
+ break;
+ }
+ }
+
+ args->structureOutputSize = structureOutputSize;
+
+ return (err);
+}
+
+
+#if __LP64__
OSMetaClassDefineReservedUnused(IOUserClient, 0);
OSMetaClassDefineReservedUnused(IOUserClient, 1);
+#else
+OSMetaClassDefineReservedUsed(IOUserClient, 0);
+OSMetaClassDefineReservedUsed(IOUserClient, 1);
+#endif
OSMetaClassDefineReservedUnused(IOUserClient, 2);
OSMetaClassDefineReservedUnused(IOUserClient, 3);
OSMetaClassDefineReservedUnused(IOUserClient, 4);