/*
- * Copyright (c) 2005 Apple Computer, Inc. All rights reserved.
- *
* @APPLE_APACHE_LICENSE_HEADER_START@
*
* Licensed under the Apache License, Version 2.0 (the "License");
* @APPLE_APACHE_LICENSE_HEADER_END@
*/
-static const char *const __rcs_file_version__ = "$Revision: 1.77 $";
+static const char *const __rcs_file_version__ = "$Revision: 24984 $";
+#include "config.h"
+#include "launchd_core_logic.h"
+#include "launch_internal.h"
+#include "launchd_helper.h"
+
+#include <TargetConditionals.h>
#include <mach/mach.h>
#include <mach/mach_error.h>
#include <mach/boolean.h>
#include <mach/host_info.h>
#include <mach/mach_host.h>
#include <mach/exception.h>
+#include <mach/host_reboot.h>
#include <sys/types.h>
#include <sys/queue.h>
#include <sys/event.h>
-#include <sys/ptrace.h>
#include <sys/stat.h>
#include <sys/ucred.h>
#include <sys/fcntl.h>
#include <sys/un.h>
+#include <sys/reboot.h>
#include <sys/wait.h>
#include <sys/sysctl.h>
#include <sys/sockio.h>
#include <sys/resource.h>
#include <sys/ioctl.h>
#include <sys/mount.h>
+#include <sys/pipe.h>
+#include <sys/mman.h>
+#include <sys/socket.h>
+#include <sys/syscall.h>
#include <net/if.h>
#include <netinet/in.h>
#include <netinet/in_var.h>
#include <netinet6/nd6.h>
+#include <bsm/libbsm.h>
#include <unistd.h>
#include <signal.h>
#include <errno.h>
-#include <syslog.h>
#include <libgen.h>
#include <stdio.h>
#include <stdlib.h>
#include <string.h>
#include <ctype.h>
#include <glob.h>
+#include <System/sys/spawn.h>
+#include <spawn.h>
+#include <time.h>
+
+#include <libproc.h>
+#include <malloc/malloc.h>
+#include <pthread.h>
+#include <libproc.h>
+#if HAVE_SANDBOX
+#define __APPLE_API_PRIVATE
+#include <sandbox.h>
+#endif
+#if HAVE_QUARANTINE
+#include <quarantine.h>
+#endif
+#if TARGET_OS_EMBEDDED
+#include <sys/kern_memorystatus.h>
+#else
+/* To make my life easier. */
+typedef struct jetsam_priority_entry {
+ pid_t pid;
+ uint32_t priority;
+ uint32_t flags;
+ int32_t hiwat_pages;
+ int32_t hiwat_reserved1;
+ int32_t hiwat_reserved2;
+ int32_t hiwat_reserved3;
+} jetsam_priority_entry_t;
+
+enum {
+ kJetsamFlagsFrontmost = (1 << 0),
+ kJetsamFlagsKilled = (1 << 1)
+};
+#endif
#include "launch.h"
#include "launch_priv.h"
+#include "launch_internal.h"
+#include "bootstrap.h"
+#include "bootstrap_priv.h"
+#include "vproc.h"
+#include "vproc_internal.h"
+
+#include "reboot2.h"
+
#include "launchd.h"
-#include "launchd_core_logic.h"
+#include "launchd_runtime.h"
#include "launchd_unix_ipc.h"
-#include "bootstrap_private.h"
-#include "bootstrap.h"
-#include "bootstrapServer.h"
-#include "mpm_reply.h"
+#include "protocol_vproc.h"
+#include "protocol_vprocServer.h"
+#include "protocol_job_reply.h"
+#include "protocol_job_forward.h"
+#include "mach_excServer.h"
+#if !TARGET_OS_EMBEDDED
+#include "domainServer.h"
+#include "init.h"
+#endif
+#include "eventsServer.h"
+
+#ifndef POSIX_SPAWN_OSX_TALAPP_START
+#define POSIX_SPAWN_OSX_TALAPP_START 0x0400
+#endif
+
+#ifndef POSIX_SPAWN_OSX_WIDGET_START
+#define POSIX_SPAWN_OSX_WIDGET_START 0x0800
+#endif
-/* <rdar://problem/2685209> sys/queue.h is not up to date */
-#ifndef SLIST_FOREACH_SAFE
-#define SLIST_FOREACH_SAFE(var, head, field, tvar) \
- for ((var) = SLIST_FIRST((head)); \
- (var) && ((tvar) = SLIST_NEXT((var), field), 1); \
- (var) = (tvar))
+#ifndef POSIX_SPAWN_IOS_APP_START
+#define POSIX_SPAWN_IOS_APP_START 0x1000
#endif
+/* LAUNCHD_DEFAULT_EXIT_TIMEOUT
+ * If the job hasn't exited in the given number of seconds after sending
+ * it a SIGTERM, SIGKILL it. Can be overriden in the job plist.
+ */
+#define LAUNCHD_MIN_JOB_RUN_TIME 10
+#define LAUNCHD_DEFAULT_EXIT_TIMEOUT 20
+#define LAUNCHD_SIGKILL_TIMER 2
+#define LAUNCHD_LOG_FAILED_EXEC_FREQ 10
+
+#define SHUTDOWN_LOG_DIR "/var/log/shutdown"
+
+#define TAKE_SUBSET_NAME "TakeSubsetName"
+#define TAKE_SUBSET_PID "TakeSubsetPID"
+#define TAKE_SUBSET_PERPID "TakeSubsetPerPID"
+
+#define IS_POWER_OF_TWO(v) (!(v & (v - 1)) && v)
+
+extern char **environ;
+
+struct waiting_for_removal {
+ SLIST_ENTRY(waiting_for_removal) sle;
+ mach_port_t reply_port;
+};
+
+static bool waiting4removal_new(job_t j, mach_port_t rp);
+static void waiting4removal_delete(job_t j, struct waiting_for_removal *w4r);
+
struct machservice {
SLIST_ENTRY(machservice) sle;
- struct jobcb *job;
+ SLIST_ENTRY(machservice) special_port_sle;
+ LIST_ENTRY(machservice) name_hash_sle;
+ LIST_ENTRY(machservice) port_hash_sle;
+ struct machservice *alias;
+ job_t job;
+ unsigned int gen_num;
mach_port_name_t port;
- unsigned int isActive:1, reset:1, recv:1, hide:1, kUNCServer:1, __junk:27;
- char name[0];
+ unsigned int
+ isActive :1,
+ reset :1,
+ recv :1,
+ hide :1,
+ kUNCServer :1,
+ per_user_hack :1,
+ debug_on_close :1,
+ per_pid :1,
+ delete_on_destruction :1,
+ drain_one_on_crash :1,
+ drain_all_on_crash :1,
+ event_update_port :1, /* The job which owns this port is the event monitor. */
+ upfront :1, /* This service was declared in the plist. */
+ event_channel :1, /* The job is to receive events on this channel. */
+ /* Don't let the size of this field to get too small. It has to be large enough
+ * to represent the reasonable range of special port numbers.
+ */
+ special_port_num :18;
+ const char name[0];
};
+static SLIST_HEAD(, machservice) special_ports; /* hack, this should be per jobmgr_t */
+
+#define PORT_HASH_SIZE 32
+#define HASH_PORT(x) (IS_POWER_OF_TWO(PORT_HASH_SIZE) ? (MACH_PORT_INDEX(x) & (PORT_HASH_SIZE - 1)) : (MACH_PORT_INDEX(x) % PORT_HASH_SIZE))
+
+static LIST_HEAD(, machservice) port_hash[PORT_HASH_SIZE];
+
static void machservice_setup(launch_data_t obj, const char *key, void *context);
static void machservice_setup_options(launch_data_t obj, const char *key, void *context);
-static void machservice_resetport(struct jobcb *j, struct machservice *ms);
-
+static void machservice_resetport(job_t j, struct machservice *ms);
+static struct machservice *machservice_new(job_t j, const char *name, mach_port_t *serviceport, bool pid_local);
+#ifndef __LAUNCH_DISABLE_XPC_SUPPORT__
+static struct machservice *machservice_new_alias(job_t aj, struct machservice *orig);
+#endif
+static void machservice_ignore(job_t j, struct machservice *ms);
+static void machservice_watch(job_t j, struct machservice *ms);
+static void machservice_delete(job_t j, struct machservice *, bool port_died);
+static void machservice_request_notifications(struct machservice *);
+static mach_port_t machservice_port(struct machservice *);
+static job_t machservice_job(struct machservice *);
+static bool machservice_hidden(struct machservice *);
+static bool machservice_active(struct machservice *);
+static const char *machservice_name(struct machservice *);
+static bootstrap_status_t machservice_status(struct machservice *);
+void machservice_drain_port(struct machservice *);
+static struct machservice *xpc_events_find_channel(job_t j, event_name_t stream, mach_port_t *p);
struct socketgroup {
SLIST_ENTRY(socketgroup) sle;
int *fds;
unsigned int junkfds:1, fd_cnt:31;
- char name[0];
+ union {
+ const char name[0];
+ char name_init[0];
+ };
};
-static bool socketgroup_new(struct jobcb *j, const char *name, int *fds, unsigned int fd_cnt, bool junkfds);
-static void socketgroup_delete(struct jobcb *j, struct socketgroup *sg);
-static void socketgroup_watch(struct jobcb *j, struct socketgroup *sg);
-static void socketgroup_ignore(struct jobcb *j, struct socketgroup *sg);
-static void socketgroup_callback(struct jobcb *j, struct kevent *kev);
+static bool socketgroup_new(job_t j, const char *name, int *fds, size_t fd_cnt, bool junkfds);
+static void socketgroup_delete(job_t j, struct socketgroup *sg);
+static void socketgroup_watch(job_t j, struct socketgroup *sg);
+static void socketgroup_ignore(job_t j, struct socketgroup *sg);
+static void socketgroup_callback(job_t j);
static void socketgroup_setup(launch_data_t obj, const char *key, void *context);
-
-struct watchpath {
- SLIST_ENTRY(watchpath) sle;
- int fd;
- unsigned int is_qdir:1, __junk:31;
- char name[0];
-};
-
-static bool watchpath_new(struct jobcb *j, const char *name, bool qdir);
-static void watchpath_delete(struct jobcb *j, struct watchpath *wp);
-static void watchpath_watch(struct jobcb *j, struct watchpath *wp);
-static void watchpath_ignore(struct jobcb *j, struct watchpath *wp);
-static void watchpath_callback(struct jobcb *j, struct kevent *kev);
+static void socketgroup_kevent_mod(job_t j, struct socketgroup *sg, bool do_add);
struct calendarinterval {
+ LIST_ENTRY(calendarinterval) global_sle;
SLIST_ENTRY(calendarinterval) sle;
+ job_t job;
struct tm when;
+ time_t when_next;
};
-static bool calendarinterval_new(struct jobcb *j, struct tm *w);
-static bool calendarinterval_new_from_obj(struct jobcb *j, launch_data_t obj);
-static void calendarinterval_delete(struct jobcb *j, struct calendarinterval *ci);
-static void calendarinterval_setalarm(struct jobcb *j, struct calendarinterval *ci);
-static void calendarinterval_callback(struct jobcb *j, struct kevent *kev);
+static LIST_HEAD(, calendarinterval) sorted_calendar_events;
+
+static bool calendarinterval_new(job_t j, struct tm *w);
+static bool calendarinterval_new_from_obj(job_t j, launch_data_t obj);
+static void calendarinterval_new_from_obj_dict_walk(launch_data_t obj, const char *key, void *context);
+static void calendarinterval_delete(job_t j, struct calendarinterval *ci);
+static void calendarinterval_setalarm(job_t j, struct calendarinterval *ci);
+static void calendarinterval_callback(void);
+static void calendarinterval_sanity_check(void);
struct envitem {
SLIST_ENTRY(envitem) sle;
+ bool one_shot;
char *value;
- char key[0];
+ union {
+ const char key[0];
+ char key_init[0];
+ };
};
-static bool envitem_new(struct jobcb *j, const char *k, const char *v, bool global);
-static void envitem_delete(struct jobcb *j, struct envitem *ei, bool global);
+static bool envitem_new(job_t j, const char *k, const char *v, bool global, bool one_shot);
+static void envitem_delete(job_t j, struct envitem *ei, bool global);
static void envitem_setup(launch_data_t obj, const char *key, void *context);
+static void envitem_setup_one_shot(launch_data_t obj, const char *key, void *context);
struct limititem {
SLIST_ENTRY(limititem) sle;
unsigned int setsoft:1, sethard:1, which:30;
};
-static bool limititem_update(struct jobcb *j, int w, rlim_t r);
-static void limititem_delete(struct jobcb *j, struct limititem *li);
+static bool limititem_update(job_t j, int w, rlim_t r);
+static void limititem_delete(job_t j, struct limititem *li);
static void limititem_setup(launch_data_t obj, const char *key, void *context);
+#if HAVE_SANDBOX
+static void seatbelt_setup_flags(launch_data_t obj, const char *key, void *context);
+#endif
+
+static void jetsam_property_setup(launch_data_t obj, const char *key, job_t j);
typedef enum {
NETWORK_UP = 1,
NETWORK_DOWN,
SUCCESSFUL_EXIT,
FAILED_EXIT,
+ CRASHED,
+ DID_NOT_CRASH,
PATH_EXISTS,
PATH_MISSING,
+ OTHER_JOB_ENABLED,
+ OTHER_JOB_DISABLED,
+ OTHER_JOB_ACTIVE,
+ OTHER_JOB_INACTIVE,
+ PATH_CHANGES,
+ DIR_NOT_EMPTY,
// FILESYSTEMTYPE_IS_MOUNTED, /* for nfsiod, but maybe others */
} semaphore_reason_t;
struct semaphoreitem {
SLIST_ENTRY(semaphoreitem) sle;
semaphore_reason_t why;
- char what[0];
+ bool watching_parent;
+ int fd;
+
+ union {
+ const char what[0];
+ char what_init[0];
+ };
+};
+
+struct semaphoreitem_dict_iter_context {
+ job_t j;
+ semaphore_reason_t why_true;
+ semaphore_reason_t why_false;
};
-static bool semaphoreitem_new(struct jobcb *j, semaphore_reason_t why, const char *what);
-static void semaphoreitem_delete(struct jobcb *j, struct semaphoreitem *si);
+static bool semaphoreitem_new(job_t j, semaphore_reason_t why, const char *what);
+static void semaphoreitem_delete(job_t j, struct semaphoreitem *si);
static void semaphoreitem_setup(launch_data_t obj, const char *key, void *context);
-static void semaphoreitem_setup_paths(launch_data_t obj, const char *key, void *context);
+static void semaphoreitem_setup_dict_iter(launch_data_t obj, const char *key, void *context);
+static void semaphoreitem_callback(job_t j, struct kevent *kev);
+static void semaphoreitem_watch(job_t j, struct semaphoreitem *si);
+static void semaphoreitem_ignore(job_t j, struct semaphoreitem *si);
+static void semaphoreitem_runtime_mod_ref(struct semaphoreitem *si, bool add);
+
+struct externalevent {
+ LIST_ENTRY(externalevent) sys_le;
+ LIST_ENTRY(externalevent) job_le;
+ struct eventsystem *sys;
+
+ uint64_t id;
+ job_t job;
+ bool state;
+ bool wanted_state;
+ launch_data_t event;
+
+ char name[0];
+};
+
+struct externalevent_iter_ctx {
+ job_t j;
+ struct eventsystem *sys;
+};
+
+static bool externalevent_new(job_t j, struct eventsystem *sys, char *evname, launch_data_t event);
+static void externalevent_delete(struct externalevent *ee);
+static void externalevent_setup(launch_data_t obj, const char *key, void *context);
+static struct externalevent *externalevent_find(const char *sysname, uint64_t id);
+
+struct eventsystem {
+ LIST_ENTRY(eventsystem) global_le;
+ LIST_HEAD(, externalevent) events;
+ uint64_t curid;
+ bool has_updates;
+ char name[0];
+};
+
+static struct eventsystem *eventsystem_new(const char *name);
+static void eventsystem_delete(struct eventsystem *sys);
+static void eventsystem_setup(launch_data_t obj, const char *key, void *context);
+static struct eventsystem *eventsystem_find(const char *name);
+static void eventsystem_ping(void);
+
+#define ACTIVE_JOB_HASH_SIZE 32
+#define ACTIVE_JOB_HASH(x) (IS_POWER_OF_TWO(ACTIVE_JOB_HASH_SIZE) ? (x & (ACTIVE_JOB_HASH_SIZE - 1)) : (x % ACTIVE_JOB_HASH_SIZE))
+
+#define MACHSERVICE_HASH_SIZE 37
+
+#define LABEL_HASH_SIZE 53
+struct jobmgr_s {
+ kq_callback kqjobmgr_callback;
+ LIST_ENTRY(jobmgr_s) xpc_le;
+ SLIST_ENTRY(jobmgr_s) sle;
+ SLIST_HEAD(, jobmgr_s) submgrs;
+ LIST_HEAD(, job_s) jobs;
+ LIST_HEAD(, job_s) jetsam_jobs;
+
+ /* For legacy reasons, we keep all job labels that are imported in the
+ * root job manager's label hash. If a job manager is an XPC domain, then
+ * it gets its own label hash that is separate from the "global" one
+ * stored in the root job manager.
+ */
+ LIST_HEAD(, job_s) label_hash[LABEL_HASH_SIZE];
+ LIST_HEAD(, job_s) active_jobs[ACTIVE_JOB_HASH_SIZE];
+ LIST_HEAD(, machservice) ms_hash[MACHSERVICE_HASH_SIZE];
+ LIST_HEAD(, job_s) global_env_jobs;
+ mach_port_t jm_port;
+ mach_port_t req_port;
+ jobmgr_t parentmgr;
+ int reboot_flags;
+ time_t shutdown_time;
+ unsigned int global_on_demand_cnt;
+ unsigned int normal_active_cnt;
+ unsigned int jetsam_jobs_cnt;
+ unsigned int
+ shutting_down :1,
+ session_initialized :1,
+ killed_stray_jobs :1,
+ monitor_shutdown :1,
+ shutdown_jobs_dirtied :1,
+ shutdown_jobs_cleaned :1,
+ xpc_singleton :1;
+ uint32_t properties;
+ /* XPC-specific properties. */
+ char owner[MAXCOMLEN];
+ char *shortdesc;
+ mach_port_t req_bsport;
+ mach_port_t req_excport;
+ mach_port_t req_asport;
+ pid_t req_pid;
+ uid_t req_euid;
+ gid_t req_egid;
+ au_asid_t req_asid;
+ vm_offset_t req_ctx;
+ mach_msg_type_number_t req_ctx_sz;
+ mach_port_t req_rport;
+ kern_return_t error;
+ union {
+ const char name[0];
+ char name_init[0];
+ };
+};
+
+/* Global XPC domains. */
+#ifndef __LAUNCH_DISABLE_XPC_SUPPORT__
+static jobmgr_t _s_xpc_system_domain;
+static LIST_HEAD(, jobmgr_s) _s_xpc_user_domains;
+static LIST_HEAD(, jobmgr_s) _s_xpc_session_domains;
+#endif
+#define jobmgr_assumes(jm, e) \
+ (unlikely(!(e)) ? jobmgr_log_bug(jm, __LINE__), false : true)
-struct jobcb {
- kq_callback kqjob_callback;
- SLIST_ENTRY(jobcb) sle;
+static jobmgr_t jobmgr_new(jobmgr_t jm, mach_port_t requestorport, mach_port_t transfer_port, bool sflag, const char *name, bool no_init, mach_port_t asport);
+#ifndef __LAUNCH_DISABLE_XPC_SUPPORT__
+static jobmgr_t jobmgr_new_xpc_singleton_domain(jobmgr_t jm, name_t name);
+static jobmgr_t jobmgr_find_xpc_per_user_domain(jobmgr_t jm, uid_t uid);
+static jobmgr_t jobmgr_find_xpc_per_session_domain(jobmgr_t jm, au_asid_t asid);
+static job_t xpc_domain_import_service(jobmgr_t jm, launch_data_t pload);
+#endif
+static job_t jobmgr_import2(jobmgr_t jm, launch_data_t pload);
+static jobmgr_t jobmgr_parent(jobmgr_t jm);
+static jobmgr_t jobmgr_do_garbage_collection(jobmgr_t jm);
+static bool jobmgr_label_test(jobmgr_t jm, const char *str);
+static void jobmgr_reap_bulk(jobmgr_t jm, struct kevent *kev);
+static void jobmgr_log_stray_children(jobmgr_t jm, bool kill_strays);
+static void jobmgr_kill_stray_children(jobmgr_t jm, pid_t *p, size_t np);
+static void jobmgr_remove(jobmgr_t jm);
+static void jobmgr_dispatch_all(jobmgr_t jm, bool newmounthack);
+static job_t jobmgr_init_session(jobmgr_t jm, const char *session_type, bool sflag);
+static job_t jobmgr_find_by_pid_deep(jobmgr_t jm, pid_t p, bool anon_okay);
+static job_t jobmgr_find_by_pid(jobmgr_t jm, pid_t p, bool create_anon);
+static jobmgr_t jobmgr_find_by_name(jobmgr_t jm, const char *where);
+static job_t job_mig_intran2(jobmgr_t jm, mach_port_t mport, pid_t upid);
+static job_t jobmgr_lookup_per_user_context_internal(job_t j, uid_t which_user, mach_port_t *mp);
+static void job_export_all2(jobmgr_t jm, launch_data_t where);
+static void jobmgr_callback(void *obj, struct kevent *kev);
+static void jobmgr_setup_env_from_other_jobs(jobmgr_t jm);
+static void jobmgr_export_env_from_other_jobs(jobmgr_t jm, launch_data_t dict);
+static struct machservice *jobmgr_lookup_service(jobmgr_t jm, const char *name, bool check_parent, pid_t target_pid);
+static void jobmgr_logv(jobmgr_t jm, int pri, int err, const char *msg, va_list ap) __attribute__((format(printf, 4, 0)));
+static void jobmgr_log(jobmgr_t jm, int pri, const char *msg, ...) __attribute__((format(printf, 3, 4)));
+/* static void jobmgr_log_error(jobmgr_t jm, int pri, const char *msg, ...) __attribute__((format(printf, 3, 4))); */
+static void jobmgr_log_bug(jobmgr_t jm, unsigned int line);
+
+#define AUTO_PICK_LEGACY_LABEL (const char *)(~0)
+#define AUTO_PICK_ANONYMOUS_LABEL (const char *)(~1)
+#define AUTO_PICK_XPC_LABEL (const char *)(~2)
+
+struct suspended_peruser {
+ LIST_ENTRY(suspended_peruser) sle;
+ job_t j;
+};
+
+struct job_s {
+ kq_callback kqjob_callback; /* MUST be first element of this structure for benefit of launchd's run loop. */
+ LIST_ENTRY(job_s) sle;
+ LIST_ENTRY(job_s) subjob_sle;
+ LIST_ENTRY(job_s) needing_session_sle;
+ LIST_ENTRY(job_s) jetsam_sle;
+ LIST_ENTRY(job_s) pid_hash_sle;
+ LIST_ENTRY(job_s) label_hash_sle;
+ LIST_ENTRY(job_s) global_env_sle;
+ SLIST_ENTRY(job_s) curious_jobs_sle;
+ LIST_HEAD(, suspended_peruser) suspended_perusers;
+ LIST_HEAD(, waiting_for_exit) exit_watchers;
+ LIST_HEAD(, job_s) subjobs;
+ LIST_HEAD(, externalevent) events;
SLIST_HEAD(, socketgroup) sockets;
- SLIST_HEAD(, watchpath) vnodes;
SLIST_HEAD(, calendarinterval) cal_intervals;
SLIST_HEAD(, envitem) global_env;
SLIST_HEAD(, envitem) env;
SLIST_HEAD(, limititem) limits;
SLIST_HEAD(, machservice) machservices;
SLIST_HEAD(, semaphoreitem) semaphores;
- SLIST_HEAD(, jobcb) jobs;
+ SLIST_HEAD(, waiting_for_removal) removal_watchers;
+ job_t alias;
struct rusage ru;
- struct jobcb *parent;
- mach_port_t bs_port;
- mach_port_t req_port;
- mach_port_t wait_reply_port;
+ cpu_type_t *j_binpref;
+ size_t j_binpref_cnt;
+ mach_port_t j_port;
+ mach_port_t exit_status_dest;
+ mach_port_t exit_status_port;
+ mach_port_t spawn_reply_port;
uid_t mach_uid;
+ jobmgr_t mgr;
+ size_t argc;
char **argv;
char *prog;
char *rootdir;
char *stdinpath;
char *stdoutpath;
char *stderrpath;
+ char *alt_exc_handler;
+ struct vproc_shmem_s *shmem;
+ struct machservice *lastlookup;
+ unsigned int lastlookup_gennum;
+#if HAVE_SANDBOX
+ char *seatbelt_profile;
+ uint64_t seatbelt_flags;
+#endif
+#if HAVE_QUARANTINE
+ void *quarantine_data;
+ size_t quarantine_data_sz;
+#endif
pid_t p;
- int argc;
int last_exit_status;
- int execfd;
+ int stdin_fd;
+ int fork_fd;
+ int log_redirect_fd;
int nice;
- int timeout;
- time_t start_time;
- size_t failed_exits;
- unsigned int start_interval;
- unsigned int checkedin:1, firstborn:1, debug:1, throttle:1, inetcompat:1, inetcompat_wait:1,
- ondemand:1, session_create:1, low_pri_io:1, init_groups:1, priv_port_has_senders:1,
- importing_global_env:1, importing_hard_limits:1, setmask:1, legacy_mach_job:1, runatload:1;
+ int stdout_err_fd;
+ uint32_t pstype;
+ int32_t jetsam_priority;
+ int32_t jetsam_memlimit;
+ int32_t jetsam_seq;
+ int32_t main_thread_priority;
+ uint32_t timeout;
+ uint32_t exit_timeout;
+ uint64_t sent_signal_time;
+ uint64_t start_time;
+ uint32_t min_run_time;
+ uint32_t start_interval;
+ uint32_t peruser_suspend_count; /* The number of jobs that have disabled this per-user launchd. */
+ uuid_t instance_id;
+ uint32_t fail_cnt;
+#if 0
+ /* someday ... */
+ enum {
+ J_TYPE_ANONYMOUS = 1,
+ J_TYPE_LANCHSERVICES,
+ J_TYPE_MACHINIT,
+ J_TYPE_INETD,
+ } j_type;
+#endif
+ bool
+ debug :1, /* man launchd.plist --> Debug */
+ ondemand :1, /* man launchd.plist --> KeepAlive == false */
+ session_create :1, /* man launchd.plist --> SessionCreate */
+ low_pri_io :1, /* man launchd.plist --> LowPriorityIO */
+ no_init_groups :1, /* man launchd.plist --> InitGroups */
+ priv_port_has_senders :1, /* a legacy mach_init concept to make bootstrap_create_server/service() work */
+ importing_global_env :1, /* a hack during job importing */
+ importing_hard_limits :1, /* a hack during job importing */
+ setmask :1, /* man launchd.plist --> Umask */
+ anonymous :1, /* a process that launchd knows about, but isn't managed by launchd */
+ checkedin :1, /* a legacy mach_init concept to detect sick jobs */
+ legacy_mach_job :1, /* a job created via bootstrap_create_server() */
+ legacy_LS_job :1, /* a job created via spawn_via_launchd() */
+ inetcompat :1, /* a legacy job that wants inetd compatible semantics */
+ inetcompat_wait :1, /* a twist on inetd compatibility */
+ start_pending :1, /* an event fired and the job should start, but not necessarily right away */
+ globargv :1, /* man launchd.plist --> EnableGlobbing */
+ wait4debugger :1, /* man launchd.plist --> WaitForDebugger */
+ wait4debugger_oneshot :1, /* One-shot WaitForDebugger. */
+ internal_exc_handler :1, /* MachExceptionHandler == true */
+ stall_before_exec :1, /* a hack to support an option of spawn_via_launchd() */
+ only_once :1, /* man launchd.plist --> LaunchOnlyOnce. Note: 5465184 Rename this to "HopefullyNeverExits" */
+ currently_ignored :1, /* Make job_ignore() / job_watch() work. If these calls were balanced, then this wouldn't be necessarily. */
+ forced_peers_to_demand_mode :1, /* A job that forced all other jobs to be temporarily launch-on-demand */
+ setnice :1, /* man launchd.plist --> Nice */
+ removal_pending :1, /* a job was asked to be unloaded/removed while running, we'll remove it after it exits */
+ sent_sigkill :1, /* job_kill() was called */
+ debug_before_kill :1, /* enter the kernel debugger before killing a job */
+ weird_bootstrap :1, /* a hack that launchd+launchctl use during jobmgr_t creation */
+ start_on_mount :1, /* man launchd.plist --> StartOnMount */
+ per_user :1, /* This job is a per-user launchd managed by the PID 1 launchd */
+ unload_at_mig_return :1, /* A job thoroughly confused launchd. We need to unload it ASAP */
+ abandon_pg :1, /* man launchd.plist --> AbandonProcessGroup */
+ ignore_pg_at_shutdown :1, /* During shutdown, do not send SIGTERM to stray processes in the process group of this job. */
+ poll_for_vfs_changes :1, /* a hack to work around the fact that kqueues don't work on all filesystems */
+ deny_job_creation :1, /* Don't let this job create new 'job_t' objects in launchd */
+ kill_via_shmem :1, /* man launchd.plist --> EnableTransactions */
+ sent_kill_via_shmem :1, /* We need to 'kill_via_shmem' once-and-only-once */
+ clean_kill :1, /* The job was sent SIGKILL because it was clean. */
+ kill_after_sample :1, /* The job is to be killed after sampling. */
+ reap_after_trace :1, /* The job exited before sample did, so we should reap it after sample is done. */
+ nosy :1, /* The job has an OtherJobEnabled KeepAlive criterion. */
+ crashed :1, /* The job is the default Mach exception handler, and it crashed. */
+ reaped :1, /* We've received NOTE_EXIT for the job. */
+ stopped :1, /* job_stop() was called. */
+ jetsam_frontmost :1, /* The job is considered "frontmost" by Jetsam. */
+ needs_kickoff :1, /* The job is to be kept alive continuously, but it must be initially kicked off. */
+ is_bootstrapper :1, /* The job is a bootstrapper. */
+ has_console :1, /* The job owns the console. */
+ embedded_special_privileges :1, /* The job runs as a non-root user on embedded but has select privileges of the root user. */
+ did_exec :1, /* The job exec(2)ed successfully. */
+ xpcproxy_did_exec :1, /* The job is an XPC service, and XPC proxy successfully exec(3)ed. */
+ holds_ref :1, /* The (anonymous) job called vprocmgr_switch_to_session(). */
+ jetsam_properties :1, /* The job has Jetsam limits in place. */
+ dedicated_instance :1, /* This job was created as the result of a look up of a service provided by a per-lookup job. */
+ multiple_instances :1, /* The job supports creating additional instances of itself. */
+ former_subjob :1, /* The sub-job was already removed from the parent's list of sub-jobs. */
+ event_monitor :1, /* The job is responsible for monitoring external events for this launchd. */
+ removing :1, /* A lame hack. */
+ disable_aslr :1, /* Disable ASLR when launching this job. */
+ xpc_service :1, /* The job is an XPC Service. */
+ shutdown_monitor :1, /* The job is the Performance team's shutdown monitor. */
+ dirty_at_shutdown :1, /* We should open a transaction for the job when shutdown begins. */
+ workaround9359725 :1; /* The job was sent SIGKILL but did not exit in a timely fashion, indicating a kernel bug. */
+
mode_t mask;
- unsigned int globargv:1, wait4debugger:1, transfer_bstrap:1, unload_at_exit:1, force_ppc:1, stall_before_exec:1, __pad:26;
- char label[0];
+ pid_t tracing_pid;
+ mach_port_t asport;
+ /* Only set for per-user launchd's. */
+ au_asid_t asid;
+ uuid_t expected_audit_uuid;
+ const char label[0];
};
-static struct jobcb *job_import2(launch_data_t pload);
+static size_t hash_label(const char *label) __attribute__((pure));
+static size_t hash_ms(const char *msstr) __attribute__((pure));
+static SLIST_HEAD(, job_s) s_curious_jobs;
+
+#define job_assumes(j, e) \
+ (unlikely(!(e)) ? job_log_bug(j, __LINE__), false : true)
+
static void job_import_keys(launch_data_t obj, const char *key, void *context);
-static void job_import_bool(struct jobcb *j, const char *key, bool value);
-static void job_import_string(struct jobcb *j, const char *key, const char *value);
-static void job_import_integer(struct jobcb *j, const char *key, long long value);
-static void job_import_dictionary(struct jobcb *j, const char *key, launch_data_t value);
-static void job_import_array(struct jobcb *j, const char *key, launch_data_t value);
-static void job_watch(struct jobcb *j);
-static void job_ignore(struct jobcb *j);
-static void job_reap(struct jobcb *j);
-static bool job_useless(struct jobcb *j);
-static bool job_keepalive(struct jobcb *j);
-static void job_start_child(struct jobcb *j, int execfd) __attribute__((noreturn));
-static void job_setup_attributes(struct jobcb *j);
-static bool job_setup_machport(struct jobcb *j);
+static void job_import_bool(job_t j, const char *key, bool value);
+static void job_import_string(job_t j, const char *key, const char *value);
+static void job_import_integer(job_t j, const char *key, long long value);
+static void job_import_dictionary(job_t j, const char *key, launch_data_t value);
+static void job_import_array(job_t j, const char *key, launch_data_t value);
+static void job_import_opaque(job_t j, const char *key, launch_data_t value);
+static bool job_set_global_on_demand(job_t j, bool val);
+static const char *job_active(job_t j);
+static void job_watch(job_t j);
+static void job_ignore(job_t j);
+static void job_cleanup_after_tracer(job_t j);
+static void job_reap(job_t j);
+static bool job_useless(job_t j);
+static bool job_keepalive(job_t j);
+static void job_dispatch_curious_jobs(job_t j);
+static void job_start(job_t j);
+static void job_start_child(job_t j) __attribute__((noreturn));
+static void job_setup_attributes(job_t j);
+static bool job_setup_machport(job_t j);
+static kern_return_t job_setup_exit_port(job_t j);
+static void job_setup_fd(job_t j, int target_fd, const char *path, int flags);
+static void job_postfork_become_user(job_t j);
+static void job_postfork_test_user(job_t j);
+static void job_log_pids_with_weird_uids(job_t j);
+static void job_setup_exception_port(job_t j, task_t target_task);
static void job_callback(void *obj, struct kevent *kev);
-static pid_t job_fork(struct jobcb *j);
-static size_t job_prep_log_preface(struct jobcb *j, char *buf);
-static void job_setup_env_from_other_jobs(struct jobcb *j);
-static void job_export_all2(struct jobcb *j, launch_data_t where);
-static launch_data_t job_export2(struct jobcb *j, bool subjobs);
-
+static void job_callback_proc(job_t j, struct kevent *kev);
+static void job_callback_timer(job_t j, void *ident);
+static void job_callback_read(job_t j, int ident);
+static void job_log_stray_pg(job_t j);
+static void job_log_children_without_exec(job_t j);
+static job_t job_new_anonymous(jobmgr_t jm, pid_t anonpid) __attribute__((malloc, nonnull, warn_unused_result));
+static job_t job_new(jobmgr_t jm, const char *label, const char *prog, const char *const *argv) __attribute__((malloc, nonnull(1,2), warn_unused_result));
+#ifndef __LAUNCH_DISABLE_XPC_SUPPORT__
+static job_t job_new_alias(jobmgr_t jm, job_t src);
+#endif
+static job_t job_new_via_mach_init(job_t j, const char *cmd, uid_t uid, bool ond) __attribute__((malloc, nonnull, warn_unused_result));
+static job_t job_new_subjob(job_t j, uuid_t identifier);
+static void job_kill(job_t j);
+static void job_uncork_fork(job_t j);
+static void job_log_stdouterr(job_t j);
+static void job_logv(job_t j, int pri, int err, const char *msg, va_list ap) __attribute__((format(printf, 4, 0)));
+static void job_log_error(job_t j, int pri, const char *msg, ...) __attribute__((format(printf, 3, 4)));
+static void job_log_bug(job_t j, unsigned int line);
+static void job_log_stdouterr2(job_t j, const char *msg, ...);
+static void job_set_exception_port(job_t j, mach_port_t port);
+static kern_return_t job_mig_spawn_internal(job_t j, vm_offset_t indata, mach_msg_type_number_t indataCnt, mach_port_t asport, job_t *outj);
+static void job_open_shutdown_transaction(job_t ji);
+static void job_close_shutdown_transaction(job_t ji);
static const struct {
const char *key;
int val;
} launchd_keys2limits[] = {
- { LAUNCH_JOBKEY_RESOURCELIMIT_CORE, RLIMIT_CORE },
- { LAUNCH_JOBKEY_RESOURCELIMIT_CPU, RLIMIT_CPU },
- { LAUNCH_JOBKEY_RESOURCELIMIT_DATA, RLIMIT_DATA },
- { LAUNCH_JOBKEY_RESOURCELIMIT_FSIZE, RLIMIT_FSIZE },
- { LAUNCH_JOBKEY_RESOURCELIMIT_MEMLOCK, RLIMIT_MEMLOCK },
- { LAUNCH_JOBKEY_RESOURCELIMIT_NOFILE, RLIMIT_NOFILE },
- { LAUNCH_JOBKEY_RESOURCELIMIT_NPROC, RLIMIT_NPROC },
- { LAUNCH_JOBKEY_RESOURCELIMIT_RSS, RLIMIT_RSS },
- { LAUNCH_JOBKEY_RESOURCELIMIT_STACK, RLIMIT_STACK },
+ { LAUNCH_JOBKEY_RESOURCELIMIT_CORE, RLIMIT_CORE },
+ { LAUNCH_JOBKEY_RESOURCELIMIT_CPU, RLIMIT_CPU },
+ { LAUNCH_JOBKEY_RESOURCELIMIT_DATA, RLIMIT_DATA },
+ { LAUNCH_JOBKEY_RESOURCELIMIT_FSIZE, RLIMIT_FSIZE },
+ { LAUNCH_JOBKEY_RESOURCELIMIT_MEMLOCK, RLIMIT_MEMLOCK },
+ { LAUNCH_JOBKEY_RESOURCELIMIT_NOFILE, RLIMIT_NOFILE },
+ { LAUNCH_JOBKEY_RESOURCELIMIT_NPROC, RLIMIT_NPROC },
+ { LAUNCH_JOBKEY_RESOURCELIMIT_RSS, RLIMIT_RSS },
+ { LAUNCH_JOBKEY_RESOURCELIMIT_STACK, RLIMIT_STACK },
};
static time_t cronemu(int mon, int mday, int hour, int min);
static bool cronemu_hour(struct tm *wtm, int hour, int min);
static bool cronemu_min(struct tm *wtm, int min);
-static void simple_zombie_reaper(void *, struct kevent *);
-
-kq_callback kqsimple_zombie_reaper = simple_zombie_reaper;
-
-static int dir_has_files(const char *path);
+/* These functions are a total nightmare to get to through headers.
+ * See rdar://problem/8223092.
+ */
+typedef __darwin_mach_port_t fileport_t;
+#define FILEPORT_NULL ((fileport_t)0)
+extern int fileport_makeport(int, fileport_t *);
+extern int fileport_makefd(fileport_t);
+
+/* miscellaneous file local functions */
+static size_t get_kern_max_proc(void);
+static int dir_has_files(job_t j, const char *path);
static char **mach_cmd2argv(const char *string);
-struct jobcb *root_job = NULL;
-struct jobcb *gc_this_job = NULL;
-size_t total_children = 0;
+static size_t our_strhash(const char *s) __attribute__((pure));
+static void extract_rcsid_substr(const char *i, char *o, size_t osz);
+
+void eliminate_double_reboot(void);
+
+/* file local globals */
+static size_t total_children;
+static size_t total_anon_children;
+static mach_port_t the_exception_server;
+static job_t workaround_5477111;
+static LIST_HEAD(, job_s) s_needing_sessions;
+static LIST_HEAD(, eventsystem) _s_event_systems;
+static job_t _s_event_monitor;
+static job_t _s_shutdown_monitor;
+static mach_port_t _s_event_update_port;
+mach_port_t g_audit_session_port = MACH_PORT_NULL;
+static uint32_t s_jetsam_sequence_id;
+
+#if !TARGET_OS_EMBEDDED
+static job_t s_embedded_privileged_job = (job_t)&root_jobmgr;
+au_asid_t g_audit_session = AU_DEFAUDITSID;
+#else
+static job_t s_embedded_privileged_job = NULL;
+pid_t g_audit_session = 0;
+#endif
-void
-simple_zombie_reaper(void *obj __attribute__((unused)), struct kevent *kev)
-{
- waitpid(kev->ident, NULL, 0);
-}
+static int s_no_hang_fd = -1;
+
+/* process wide globals */
+mach_port_t inherited_bootstrap_port;
+jobmgr_t root_jobmgr;
+bool g_shutdown_debugging = false;
+bool g_verbose_boot = false;
+bool g_embedded_privileged_action = false;
+bool g_runtime_busy_time = false;
void
-job_ignore(struct jobcb *j)
+job_ignore(job_t j)
{
+ struct semaphoreitem *si;
struct socketgroup *sg;
struct machservice *ms;
- struct watchpath *wp;
- SLIST_FOREACH(sg, &j->sockets, sle)
+ if (j->currently_ignored) {
+ return;
+ }
+
+ job_log(j, LOG_DEBUG, "Ignoring...");
+
+ j->currently_ignored = true;
+
+ if (j->poll_for_vfs_changes) {
+ j->poll_for_vfs_changes = false;
+ (void)job_assumes(j, kevent_mod((uintptr_t)&j->semaphores, EVFILT_TIMER, EV_DELETE, 0, 0, j) != -1);
+ }
+
+ SLIST_FOREACH(sg, &j->sockets, sle) {
socketgroup_ignore(j, sg);
+ }
- SLIST_FOREACH(wp, &j->vnodes, sle)
- watchpath_ignore(j, wp);
+ SLIST_FOREACH(ms, &j->machservices, sle) {
+ machservice_ignore(j, ms);
+ }
- SLIST_FOREACH(ms, &j->machservices, sle)
- launchd_assumes(launchd_mport_request_callback(ms->port, NULL, false) == KERN_SUCCESS);
+ SLIST_FOREACH(si, &j->semaphores, sle) {
+ semaphoreitem_ignore(j, si);
+ }
}
void
-job_watch(struct jobcb *j)
+job_watch(job_t j)
{
+ struct semaphoreitem *si;
struct socketgroup *sg;
struct machservice *ms;
- struct watchpath *wp;
- SLIST_FOREACH(sg, &j->sockets, sle)
+ if (!j->currently_ignored) {
+ return;
+ }
+
+ job_log(j, LOG_DEBUG, "Watching...");
+
+ j->currently_ignored = false;
+
+ SLIST_FOREACH(sg, &j->sockets, sle) {
socketgroup_watch(j, sg);
+ }
- SLIST_FOREACH(wp, &j->vnodes, sle)
- watchpath_watch(j, wp);
+ SLIST_FOREACH(ms, &j->machservices, sle) {
+ machservice_watch(j, ms);
+ }
- SLIST_FOREACH(ms, &j->machservices, sle)
- launchd_assumes(launchd_mport_request_callback(ms->port, j, false) == KERN_SUCCESS);
+ SLIST_FOREACH(si, &j->semaphores, sle) {
+ semaphoreitem_watch(j, si);
+ }
}
void
-job_stop(struct jobcb *j)
+job_stop(job_t j)
{
- if (j->p)
- kill(j->p, SIGTERM);
-}
+ char extralog[100];
+ int32_t newval = 1;
-launch_data_t
-job_export(struct jobcb *j)
-{
- return job_export2(j, true);
+ if (unlikely(!j->p || j->stopped || j->anonymous)) {
+ return;
+ }
+
+#if TARGET_OS_EMBEDDED
+ if (g_embedded_privileged_action && s_embedded_privileged_job) {
+ if (!job_assumes(j, s_embedded_privileged_job->username != NULL && j->username != NULL)) {
+ errno = EPERM;
+ return;
+ }
+
+ if (strcmp(j->username, s_embedded_privileged_job->username) != 0) {
+ errno = EPERM;
+ return;
+ }
+ } else if (g_embedded_privileged_action) {
+ errno = EINVAL;
+ return;
+ }
+#endif
+
+ if (j->kill_via_shmem) {
+ if (j->shmem) {
+ if (!j->sent_kill_via_shmem) {
+ j->shmem->vp_shmem_flags |= VPROC_SHMEM_EXITING;
+ newval = __sync_sub_and_fetch(&j->shmem->vp_shmem_transaction_cnt, 1);
+ j->sent_kill_via_shmem = true;
+ } else {
+ newval = j->shmem->vp_shmem_transaction_cnt;
+ }
+ } else {
+ newval = -1;
+ }
+ }
+
+ j->sent_signal_time = runtime_get_opaque_time();
+
+ if (newval < 0) {
+ j->clean_kill = true;
+ job_kill(j);
+ } else {
+ (void)job_assumes(j, runtime_kill(j->p, SIGTERM) != -1);
+
+ if (j->exit_timeout) {
+ (void)job_assumes(j, kevent_mod((uintptr_t)&j->exit_timeout, EVFILT_TIMER, EV_ADD|EV_ONESHOT, NOTE_SECONDS, j->exit_timeout, j) != -1);
+ } else {
+ job_log(j, LOG_NOTICE, "This job has an infinite exit timeout");
+ }
+
+ if (j->kill_via_shmem) {
+ snprintf(extralog, sizeof(extralog), ": %d remaining transactions", newval + 1);
+ } else {
+ extralog[0] = '\0';
+ }
+
+ job_log(j, LOG_DEBUG, "Sent SIGTERM signal%s", extralog);
+ }
+
+ j->stopped = true;
}
launch_data_t
-job_export2(struct jobcb *j, bool subjobs)
+job_export(job_t j)
{
launch_data_t tmp, tmp2, tmp3, r = launch_data_alloc(LAUNCH_DATA_DICTIONARY);
- if (r == NULL)
+ if (r == NULL) {
return NULL;
+ }
- if ((tmp = launch_data_new_string(j->label)))
+ if ((tmp = launch_data_new_string(j->label))) {
launch_data_dict_insert(r, tmp, LAUNCH_JOBKEY_LABEL);
-
- if ((tmp = launch_data_new_bool(j->ondemand)))
+ }
+ if ((tmp = launch_data_new_string(j->mgr->name))) {
+ launch_data_dict_insert(r, tmp, LAUNCH_JOBKEY_LIMITLOADTOSESSIONTYPE);
+ }
+ if ((tmp = launch_data_new_bool(j->ondemand))) {
launch_data_dict_insert(r, tmp, LAUNCH_JOBKEY_ONDEMAND);
-
- if ((tmp = launch_data_new_integer(j->last_exit_status)))
+ }
+ if ((tmp = launch_data_new_integer(j->last_exit_status))) {
launch_data_dict_insert(r, tmp, LAUNCH_JOBKEY_LASTEXITSTATUS);
-
- if (j->p && (tmp = launch_data_new_integer(j->p)))
+ }
+ if (j->p && (tmp = launch_data_new_integer(j->p))) {
launch_data_dict_insert(r, tmp, LAUNCH_JOBKEY_PID);
-
- if ((tmp = launch_data_new_integer(j->timeout)))
+ }
+ if ((tmp = launch_data_new_integer(j->timeout))) {
launch_data_dict_insert(r, tmp, LAUNCH_JOBKEY_TIMEOUT);
-
- if (j->prog && (tmp = launch_data_new_string(j->prog)))
+ }
+ if (j->prog && (tmp = launch_data_new_string(j->prog))) {
launch_data_dict_insert(r, tmp, LAUNCH_JOBKEY_PROGRAM);
-
- if (j->stdoutpath && (tmp = launch_data_new_string(j->stdoutpath)))
+ }
+ if (j->stdinpath && (tmp = launch_data_new_string(j->stdinpath))) {
+ launch_data_dict_insert(r, tmp, LAUNCH_JOBKEY_STANDARDINPATH);
+ }
+ if (j->stdoutpath && (tmp = launch_data_new_string(j->stdoutpath))) {
launch_data_dict_insert(r, tmp, LAUNCH_JOBKEY_STANDARDOUTPATH);
-
- if (j->stderrpath && (tmp = launch_data_new_string(j->stderrpath)))
+ }
+ if (j->stderrpath && (tmp = launch_data_new_string(j->stderrpath))) {
launch_data_dict_insert(r, tmp, LAUNCH_JOBKEY_STANDARDERRORPATH);
-
- if (j->argv && (tmp = launch_data_alloc(LAUNCH_DATA_ARRAY))) {
- int i;
+ }
+ if (likely(j->argv) && (tmp = launch_data_alloc(LAUNCH_DATA_ARRAY))) {
+ size_t i;
for (i = 0; i < j->argc; i++) {
- if ((tmp2 = launch_data_new_string(j->argv[i])))
+ if ((tmp2 = launch_data_new_string(j->argv[i]))) {
launch_data_array_set_index(tmp, tmp2, i);
+ }
}
launch_data_dict_insert(r, tmp, LAUNCH_JOBKEY_PROGRAMARGUMENTS);
}
+ if (j->kill_via_shmem && (tmp = launch_data_new_bool(true))) {
+ int32_t tmp_cnt = -1;
+
+ launch_data_dict_insert(r, tmp, LAUNCH_JOBKEY_ENABLETRANSACTIONS);
+
+ if (j->shmem) {
+ tmp_cnt = j->shmem->vp_shmem_transaction_cnt;
+ }
+
+ if (j->sent_kill_via_shmem) {
+ tmp_cnt++;
+ }
+
+ if ((tmp = launch_data_new_integer(tmp_cnt))) {
+ launch_data_dict_insert(r, tmp, LAUNCH_JOBKEY_TRANSACTIONCOUNT);
+ }
+ }
+
+ if (j->session_create && (tmp = launch_data_new_bool(true))) {
+ launch_data_dict_insert(r, tmp, LAUNCH_JOBKEY_SESSIONCREATE);
+ }
+
if (j->inetcompat && (tmp = launch_data_alloc(LAUNCH_DATA_DICTIONARY))) {
- if ((tmp2 = launch_data_new_bool(j->inetcompat_wait)))
+ if ((tmp2 = launch_data_new_bool(j->inetcompat_wait))) {
launch_data_dict_insert(tmp, tmp2, LAUNCH_JOBINETDCOMPATIBILITY_WAIT);
+ }
launch_data_dict_insert(r, tmp, LAUNCH_JOBKEY_INETDCOMPATIBILITY);
}
int i;
SLIST_FOREACH(sg, &j->sockets, sle) {
+ if (sg->junkfds) {
+ continue;
+ }
if ((tmp2 = launch_data_alloc(LAUNCH_DATA_ARRAY))) {
for (i = 0; i < sg->fd_cnt; i++) {
- if ((tmp3 = launch_data_new_fd(sg->fds[i])))
+ if ((tmp3 = launch_data_new_fd(sg->fds[i]))) {
launch_data_array_set_index(tmp2, tmp3, i);
+ }
}
launch_data_dict_insert(tmp, tmp2, sg->name);
}
if (!SLIST_EMPTY(&j->machservices) && (tmp = launch_data_alloc(LAUNCH_DATA_DICTIONARY))) {
struct machservice *ms;
+
+ tmp3 = NULL;
SLIST_FOREACH(ms, &j->machservices, sle) {
- tmp2 = launch_data_new_machport(MACH_PORT_NULL);
- launch_data_dict_insert(tmp, tmp2, ms->name);
+ if (ms->per_pid) {
+ if (tmp3 == NULL) {
+ tmp3 = launch_data_alloc(LAUNCH_DATA_DICTIONARY);
+ }
+ if (tmp3) {
+ tmp2 = launch_data_new_machport(MACH_PORT_NULL);
+ launch_data_dict_insert(tmp3, tmp2, ms->name);
+ }
+ } else {
+ tmp2 = launch_data_new_machport(MACH_PORT_NULL);
+ launch_data_dict_insert(tmp, tmp2, ms->name);
+ }
}
launch_data_dict_insert(r, tmp, LAUNCH_JOBKEY_MACHSERVICES);
- }
-
- if (subjobs && !SLIST_EMPTY(&j->jobs) && (tmp = launch_data_alloc(LAUNCH_DATA_ARRAY))) {
- struct jobcb *ji;
- size_t i = 0;
- SLIST_FOREACH(ji, &j->jobs, sle) {
- tmp2 = job_export2(ji, true);
- launch_data_array_set_index(tmp, tmp2, i);
- i++;
+ if (tmp3) {
+ launch_data_dict_insert(r, tmp3, LAUNCH_JOBKEY_PERJOBMACHSERVICES);
}
-
- launch_data_dict_insert(r, tmp, LAUNCH_JOBKEY_SUBJOBS);
}
return r;
}
-void
-job_remove_all_inactive(struct jobcb *j)
+static void
+jobmgr_log_active_jobs(jobmgr_t jm)
{
- struct jobcb *ji;
+ const char *why_active;
+ jobmgr_t jmi;
+ job_t ji;
- SLIST_FOREACH(ji, &j->jobs, sle)
- job_remove_all_inactive(ji);
+ SLIST_FOREACH(jmi, &jm->submgrs, sle) {
+ jobmgr_log_active_jobs(jmi);
+ }
- if (!job_active(j)) {
- job_remove(j);
- } else if (getpid() != 1) {
- job_stop(j);
+ LIST_FOREACH(ji, &jm->jobs, sle) {
+ if ((why_active = job_active(ji))) {
+ if (ji->p != 1) {
+ job_log(ji, LOG_DEBUG | LOG_CONSOLE, "%s", why_active);
+ }
+ }
}
}
-void
-job_remove(struct jobcb *j)
+static void
+jobmgr_still_alive_with_check(jobmgr_t jm)
{
- struct jobcb *ji;
- struct calendarinterval *ci;
- struct socketgroup *sg;
- struct watchpath *wp;
- struct limititem *li;
- struct envitem *ei;
- struct machservice *ms;
- struct semaphoreitem *si;
+ jobmgr_log(jm, LOG_DEBUG | LOG_CONSOLE, "Still alive with %lu/%lu (normal/anonymous) children.", total_children, total_anon_children);
+ jobmgr_log_active_jobs(jm);
+}
- job_log(j, LOG_DEBUG, "Removed");
+jobmgr_t
+jobmgr_shutdown(jobmgr_t jm)
+{
+ jobmgr_t jmi, jmn;
+ jobmgr_log(jm, LOG_DEBUG, "Beginning job manager shutdown with flags: %s", reboot_flags_to_C_names(jm->reboot_flags));
- if (j->p) {
- if (kevent_mod(j->p, EVFILT_PROC, EV_ADD, NOTE_EXIT, 0, &kqsimple_zombie_reaper) == -1) {
- job_reap(j);
- } else {
- /* we've attached the simple zombie reaper, we're going to delete the job before it is dead */
- total_children--;
- job_stop(j);
- }
- }
+ jm->shutdown_time = runtime_get_wall_time() / USEC_PER_SEC;
- if (j->parent)
- SLIST_REMOVE(&j->parent->jobs, j, jobcb, sle);
+ struct tm curtime;
+ (void)localtime_r(&jm->shutdown_time, &curtime);
- if (j->execfd)
- launchd_assumes(close(j->execfd) == 0);
+ char date[26];
+ (void)asctime_r(&curtime, date);
+ /* Trim the new line that asctime_r(3) puts there for some reason. */
+ date[24] = 0;
- if (j->bs_port) {
- if (j->transfer_bstrap) {
- launchd_assumes(launchd_mport_deallocate(j->bs_port) == KERN_SUCCESS);
- } else {
- launchd_assumes(launchd_mport_close_recv(j->bs_port) == KERN_SUCCESS);
- }
+ if (jm == root_jobmgr && pid1_magic) {
+ jobmgr_log(jm, LOG_DEBUG | LOG_CONSOLE, "Userspace shutdown begun at: %s", date);
+ } else {
+ jobmgr_log(jm, LOG_DEBUG, "Job manager shutdown begun at: %s", date);
}
- if (j->req_port)
- launchd_assumes(launchd_mport_deallocate(j->req_port) == KERN_SUCCESS);
+ jm->shutting_down = true;
-#if 0
- if (j->wait_reply_port) {
+ SLIST_FOREACH_SAFE(jmi, &jm->submgrs, sle, jmn) {
+ jobmgr_shutdown(jmi);
}
-#endif
-
- while ((ji = SLIST_FIRST(&j->jobs)))
- job_remove(ji);
+
+ if (jm->parentmgr == NULL && pid1_magic) {
+ (void)jobmgr_assumes(jm, kevent_mod((uintptr_t)jm, EVFILT_TIMER, EV_ADD, NOTE_SECONDS, 5, jm));
- while ((sg = SLIST_FIRST(&j->sockets)))
- socketgroup_delete(j, sg);
+ /* Spawn the shutdown monitor. */
+ if (_s_shutdown_monitor && !_s_shutdown_monitor->p) {
+ job_log(_s_shutdown_monitor, LOG_NOTICE | LOG_CONSOLE, "Starting shutdown monitor.");
+ job_dispatch(_s_shutdown_monitor, true);
+ }
+ }
- while ((wp = SLIST_FIRST(&j->vnodes)))
- watchpath_delete(j, wp);
+ return jobmgr_do_garbage_collection(jm);
+}
- while ((ci = SLIST_FIRST(&j->cal_intervals)))
- calendarinterval_delete(j, ci);
+void
+jobmgr_remove(jobmgr_t jm)
+{
+ jobmgr_t jmi;
+ job_t ji;
- while ((ei = SLIST_FIRST(&j->env)))
- envitem_delete(j, ei, false);
+ jobmgr_log(jm, LOG_DEBUG, "Removing job manager.");
+ if (!jobmgr_assumes(jm, SLIST_EMPTY(&jm->submgrs))) {
+ while ((jmi = SLIST_FIRST(&jm->submgrs))) {
+ jobmgr_remove(jmi);
+ }
+ }
- while ((ei = SLIST_FIRST(&j->global_env)))
- envitem_delete(j, ei, true);
+ while ((ji = LIST_FIRST(&jm->jobs))) {
+ if (!ji->anonymous && !job_assumes(ji, ji->p == 0)) {
+ ji->p = 0;
+ }
+ job_remove(ji);
+ }
- while ((li = SLIST_FIRST(&j->limits)))
- limititem_delete(j, li);
+ if (jm->req_port) {
+ (void)jobmgr_assumes(jm, launchd_mport_deallocate(jm->req_port) == KERN_SUCCESS);
+ }
+ if (jm->jm_port) {
+ (void)jobmgr_assumes(jm, launchd_mport_close_recv(jm->jm_port) == KERN_SUCCESS);
+ }
- while ((ms = SLIST_FIRST(&j->machservices)))
- machservice_delete(ms);
+ if (jm->req_bsport) {
+ (void)jobmgr_assumes(jm, launchd_mport_deallocate(jm->req_bsport) == KERN_SUCCESS);
+ }
+ if (jm->req_excport) {
+ (void)jobmgr_assumes(jm, launchd_mport_deallocate(jm->req_excport) == KERN_SUCCESS);
+ }
+ if (jm->req_asport) {
+ (void)jobmgr_assumes(jm, launchd_mport_deallocate(jm->req_asport) == KERN_SUCCESS);
+ }
+#if !TARGET_OS_EMBEDDED
+ if (jm->req_rport) {
+ kern_return_t kr = xpc_call_wakeup(jm->req_rport, jm->error);
+ if (!(kr == KERN_SUCCESS || kr == MACH_SEND_INVALID_DEST)) {
+ /* If the originator went away, the reply port will be a dead name,
+ * and we expect this to fail.
+ */
+ errno = kr;
+ (void)jobmgr_assumes(jm, kr == KERN_SUCCESS);
+ }
+ }
+#endif
+ if (jm->req_ctx) {
+ (void)jobmgr_assumes(jm, vm_deallocate(mach_task_self(), jm->req_ctx, jm->req_ctx_sz) == KERN_SUCCESS);
+ }
- while ((si = SLIST_FIRST(&j->semaphores)))
- semaphoreitem_delete(j, si);
+ time_t ts = runtime_get_wall_time() / USEC_PER_SEC;
+ struct tm curtime;
+ (void)localtime_r(&ts, &curtime);
- if (j->prog)
- free(j->prog);
+ char date[26];
+ (void)asctime_r(&curtime, date);
+ date[24] = 0;
- if (j->argv)
- free(j->argv);
+ time_t delta = ts - jm->shutdown_time;
+ if (jm == root_jobmgr && pid1_magic) {
+ jobmgr_log(jm, LOG_DEBUG | LOG_CONSOLE, "Userspace shutdown finished at: %s", date);
+ jobmgr_log(jm, LOG_DEBUG | LOG_CONSOLE, "Userspace shutdown took approximately %ld second%s.", delta, (delta != 1) ? "s" : "");
+ } else {
+ jobmgr_log(jm, LOG_DEBUG, "Job manager shutdown finished at: %s", date);
+ jobmgr_log(jm, LOG_DEBUG, "Job manager shutdown took approximately %ld second%s.", delta, (delta != 1) ? "s" : "");
+ }
- if (j->rootdir)
- free(j->rootdir);
+ if (jm->parentmgr) {
+ runtime_del_weak_ref();
+ SLIST_REMOVE(&jm->parentmgr->submgrs, jm, jobmgr_s, sle);
+ } else if (pid1_magic) {
+ eliminate_double_reboot();
+ launchd_log_vm_stats();
+ jobmgr_log_stray_children(jm, true);
+ jobmgr_log(root_jobmgr, LOG_NOTICE | LOG_CONSOLE, "About to call: reboot(%s).", reboot_flags_to_C_names(jm->reboot_flags));
+ runtime_closelog();
+ (void)jobmgr_assumes(jm, reboot(jm->reboot_flags) != -1);
+ } else {
+ jobmgr_log(jm, LOG_DEBUG, "About to exit");
+ runtime_closelog();
+ exit(EXIT_SUCCESS);
+ }
+
+ free(jm);
+}
- if (j->workingdir)
- free(j->workingdir);
+void
+job_remove(job_t j)
+{
+ struct waiting_for_removal *w4r;
+ struct calendarinterval *ci;
+ struct semaphoreitem *si;
+ struct socketgroup *sg;
+ struct machservice *ms;
+ struct limititem *li;
+ struct envitem *ei;
- if (j->username)
- free(j->username);
+ if (j->alias) {
+ /* HACK: Egregious code duplication. But as with machservice_delete(),
+ * job aliases can't (and shouldn't) have any complex behaviors
+ * associated with them.
+ */
+ while ((ms = SLIST_FIRST(&j->machservices))) {
+ machservice_delete(j, ms, false);
+ }
- if (j->groupname)
- free(j->groupname);
+ LIST_REMOVE(j, sle);
+ LIST_REMOVE(j, label_hash_sle);
+ free(j);
+ return;
+ }
- if (j->stdinpath)
- free(j->stdinpath);
+#if TARGET_OS_EMBEDDED
+ if (g_embedded_privileged_action && s_embedded_privileged_job) {
+ if (!job_assumes(j, s_embedded_privileged_job->username != NULL && j->username != NULL)) {
+ errno = EPERM;
+ return;
+ }
+
+ if (strcmp(j->username, s_embedded_privileged_job->username) != 0) {
+ errno = EPERM;
+ return;
+ }
+ } else if (g_embedded_privileged_action) {
+ errno = EINVAL;
+ return;
+ }
+#endif
+
+ /* Do this BEFORE we check and see whether the job is still active. If we're a
+ * sub-job, we're being removed due to the parent job removing us. Therefore, the
+ * parent job will free itself after this call completes. So if we defer removing
+ * ourselves from the parent's list, we'll crash when we finally get around to it.
+ */
+ if (j->dedicated_instance && !j->former_subjob) {
+ LIST_REMOVE(j, subjob_sle);
+ j->former_subjob = true;
+ }
+
+ if (unlikely(j->p)) {
+ if (j->anonymous) {
+ job_reap(j);
+ } else {
+ job_log(j, LOG_DEBUG, "Removal pended until the job exits");
- if (j->stdoutpath)
- free(j->stdoutpath);
+ if (!j->removal_pending) {
+ j->removal_pending = true;
+ job_stop(j);
+ }
+
+ return;
+ }
+ }
+
+ if (!j->removing) {
+ j->removing = true;
+ job_dispatch_curious_jobs(j);
+ }
- if (j->stderrpath)
- free(j->stderrpath);
+ ipc_close_all_with_job(j);
+
+ job_log(j, LOG_INFO, "Total rusage: utime %ld.%06u stime %ld.%06u maxrss %lu ixrss %lu idrss %lu isrss %lu minflt %lu majflt %lu nswap %lu inblock %lu oublock %lu msgsnd %lu msgrcv %lu nsignals %lu nvcsw %lu nivcsw %lu",
+ j->ru.ru_utime.tv_sec, j->ru.ru_utime.tv_usec,
+ j->ru.ru_stime.tv_sec, j->ru.ru_stime.tv_usec,
+ j->ru.ru_maxrss, j->ru.ru_ixrss, j->ru.ru_idrss, j->ru.ru_isrss,
+ j->ru.ru_minflt, j->ru.ru_majflt,
+ j->ru.ru_nswap, j->ru.ru_inblock, j->ru.ru_oublock,
+ j->ru.ru_msgsnd, j->ru.ru_msgrcv,
+ j->ru.ru_nsignals, j->ru.ru_nvcsw, j->ru.ru_nivcsw);
+
+ if (j->forced_peers_to_demand_mode) {
+ job_set_global_on_demand(j, false);
+ }
+
+ if (!job_assumes(j, j->fork_fd == 0)) {
+ (void)job_assumes(j, runtime_close(j->fork_fd) != -1);
+ }
- if (j->start_interval)
- kevent_mod((uintptr_t)&j->start_interval, EVFILT_TIMER, EV_DELETE, 0, 0, NULL);
+ if (j->stdin_fd) {
+ (void)job_assumes(j, runtime_close(j->stdin_fd) != -1);
+ }
+
+ if (!job_assumes(j, j->log_redirect_fd == 0)) {
+ (void)job_assumes(j, runtime_close(j->log_redirect_fd) != -1);
+ }
+
+ if (j->j_port) {
+ (void)job_assumes(j, launchd_mport_close_recv(j->j_port) == KERN_SUCCESS);
+ }
+
+ while ((sg = SLIST_FIRST(&j->sockets))) {
+ socketgroup_delete(j, sg);
+ }
+ while ((ci = SLIST_FIRST(&j->cal_intervals))) {
+ calendarinterval_delete(j, ci);
+ }
+ while ((ei = SLIST_FIRST(&j->env))) {
+ envitem_delete(j, ei, false);
+ }
+ while ((ei = SLIST_FIRST(&j->global_env))) {
+ envitem_delete(j, ei, true);
+ }
+ while ((li = SLIST_FIRST(&j->limits))) {
+ limititem_delete(j, li);
+ }
+ while ((ms = SLIST_FIRST(&j->machservices))) {
+ machservice_delete(j, ms, false);
+ }
+ while ((si = SLIST_FIRST(&j->semaphores))) {
+ semaphoreitem_delete(j, si);
+ }
+ while ((w4r = SLIST_FIRST(&j->removal_watchers))) {
+ waiting4removal_delete(j, w4r);
+ }
+
+ struct externalevent *eei = NULL;
+ while ((eei = LIST_FIRST(&j->events))) {
+ eventsystem_ping();
+ externalevent_delete(eei);
+ }
+
+#if 0
+ /* Event systems exist independently of an actual monitor job. They're
+ * created on-demand when a job has a LaunchEvents dictionary. So we
+ * really don't need to get rid of them.
+ */
+ if (j->event_monitor) {
+ struct eventsystem *esi = NULL;
+ while ((esi = LIST_FIRST(&_s_event_systems))) {
+ eventsystem_delete(esi);
+ }
+ }
+#else
+ if (false) {
+ /* Make gcc happy. */
+ eventsystem_delete(NULL);
+ }
+ if (j->event_monitor) {
+ if (_s_event_update_port != MACH_PORT_NULL) {
+ (void)job_assumes(j, launchd_mport_deallocate(_s_event_update_port) == KERN_SUCCESS);
+ _s_event_update_port = MACH_PORT_NULL;
+ }
+ _s_event_monitor = NULL;
+ }
+#endif
+ if (j->prog) {
+ free(j->prog);
+ }
+ if (j->argv) {
+ free(j->argv);
+ }
+ if (j->rootdir) {
+ free(j->rootdir);
+ }
+ if (j->workingdir) {
+ free(j->workingdir);
+ }
+ if (j->username) {
+ free(j->username);
+ }
+ if (j->groupname) {
+ free(j->groupname);
+ }
+ if (j->stdinpath) {
+ free(j->stdinpath);
+ }
+ if (j->stdoutpath) {
+ free(j->stdoutpath);
+ }
+ if (j->stderrpath) {
+ free(j->stderrpath);
+ }
+ if (j->alt_exc_handler) {
+ free(j->alt_exc_handler);
+ }
+#if HAVE_SANDBOX
+ if (j->seatbelt_profile) {
+ free(j->seatbelt_profile);
+ }
+#endif
+#if HAVE_QUARANTINE
+ if (j->quarantine_data) {
+ free(j->quarantine_data);
+ }
+#endif
+ if (j->j_binpref) {
+ free(j->j_binpref);
+ }
+ if (j->start_interval) {
+ runtime_del_weak_ref();
+ (void)job_assumes(j, kevent_mod((uintptr_t)&j->start_interval, EVFILT_TIMER, EV_DELETE, 0, 0, NULL) != -1);
+ }
+ if (j->poll_for_vfs_changes) {
+ (void)job_assumes(j, kevent_mod((uintptr_t)&j->semaphores, EVFILT_TIMER, EV_DELETE, 0, 0, j) != -1);
+ }
+ if (j->exit_timeout) {
+ /* Not a big deal if this fails. It means that the timer's already been freed. */
+ kevent_mod((uintptr_t)&j->exit_timeout, EVFILT_TIMER, EV_DELETE, 0, 0, NULL);
+ }
+ if (j->jetsam_properties) {
+ LIST_REMOVE(j, jetsam_sle);
+ j->mgr->jetsam_jobs_cnt--;
+ }
+ if (j->asport != MACH_PORT_NULL) {
+ (void)job_assumes(j, launchd_mport_deallocate(j->asport) == KERN_SUCCESS);
+ }
+ if (!uuid_is_null(j->expected_audit_uuid)) {
+ LIST_REMOVE(j, needing_session_sle);
+ }
+ if (j->embedded_special_privileges) {
+ s_embedded_privileged_job = NULL;
+ }
+ if (j->shutdown_monitor) {
+ _s_shutdown_monitor = NULL;
+ }
+ if (j->workaround9359725) {
+ /* We may have forcibly removed this job by simulating an exit. If this
+ * is the case, we don't want to hear about these events anymore, lest
+ * we get a stale context pointer and crash trying to dereference it.
+ */
+ kevent_mod((uintptr_t)j->p, EVFILT_PROC, EV_DELETE, 0, 0, NULL);
+ }
+
kevent_mod((uintptr_t)j, EVFILT_TIMER, EV_DELETE, 0, 0, NULL);
+
+ LIST_REMOVE(j, sle);
+ LIST_REMOVE(j, label_hash_sle);
+
+ job_t ji = NULL;
+ job_t jit = NULL;
+ LIST_FOREACH_SAFE(ji, &j->subjobs, subjob_sle, jit) {
+ job_remove(ji);
+ }
+
+ job_log(j, LOG_DEBUG, "Removed");
+
free(j);
}
socketgroup_setup(launch_data_t obj, const char *key, void *context)
{
launch_data_t tmp_oai;
- struct jobcb *j = context;
- unsigned int i, fd_cnt = 1;
+ job_t j = context;
+ size_t i, fd_cnt = 1;
int *fds;
- if (launch_data_get_type(obj) == LAUNCH_DATA_ARRAY)
+ if (launch_data_get_type(obj) == LAUNCH_DATA_ARRAY) {
fd_cnt = launch_data_array_get_count(obj);
+ }
fds = alloca(fd_cnt * sizeof(int));
for (i = 0; i < fd_cnt; i++) {
- if (launch_data_get_type(obj) == LAUNCH_DATA_ARRAY)
+ if (launch_data_get_type(obj) == LAUNCH_DATA_ARRAY) {
tmp_oai = launch_data_array_get_index(obj, i);
- else
+ } else {
tmp_oai = obj;
+ }
fds[i] = launch_data_get_fd(tmp_oai);
}
}
bool
-job_setup_machport(struct jobcb *j)
+job_set_global_on_demand(job_t j, bool val)
{
- if (!launchd_assumes(launchd_mport_create_recv(&j->bs_port) == KERN_SUCCESS))
- goto out_bad;
+ if (j->forced_peers_to_demand_mode && val) {
+ return false;
+ } else if (!j->forced_peers_to_demand_mode && !val) {
+ return false;
+ }
- if (!launchd_assumes(launchd_mport_request_callback(j->bs_port, j, true) == KERN_SUCCESS))
- goto out_bad2;
+ if ((j->forced_peers_to_demand_mode = val)) {
+ j->mgr->global_on_demand_cnt++;
+ } else {
+ j->mgr->global_on_demand_cnt--;
+ }
+
+ if (j->mgr->global_on_demand_cnt == 0) {
+ jobmgr_dispatch_all(j->mgr, false);
+ }
return true;
-out_bad2:
- launchd_assumes(launchd_mport_close_recv(j->bs_port) == KERN_SUCCESS);
-out_bad:
- return false;
}
-struct jobcb *
-job_new_via_mach_init(struct jobcb *jbs, const char *cmd, uid_t uid, bool ond)
+bool
+job_setup_machport(job_t j)
{
- const char **argv = (const char **)mach_cmd2argv(cmd);
- struct jobcb *j = NULL;
- char buf[1000];
-
- if (!launchd_assumes(argv != NULL))
- goto out_bad;
-
- /* preflight the string so we know how big it is */
- sprintf(buf, "100000.%s", basename((char *)argv[0]));
-
- j = job_new(jbs, buf, NULL, argv, NULL, MACH_PORT_NULL);
-
- free(argv);
+ mach_msg_size_t mxmsgsz;
- if (!launchd_assumes(j != NULL))
+ if (!job_assumes(j, launchd_mport_create_recv(&j->j_port) == KERN_SUCCESS)) {
goto out_bad;
+ }
- j->mach_uid = uid;
- j->ondemand = ond;
- j->legacy_mach_job = true;
- j->priv_port_has_senders = true; /* the IPC that called us will make-send on this port */
+ /* Sigh... at the moment, MIG has maxsize == sizeof(reply union) */
+ mxmsgsz = (typeof(mxmsgsz)) sizeof(union __RequestUnion__job_mig_protocol_vproc_subsystem);
+ if (job_mig_protocol_vproc_subsystem.maxsize > mxmsgsz) {
+ mxmsgsz = job_mig_protocol_vproc_subsystem.maxsize;
+ }
- if (!job_setup_machport(j))
- goto out_bad;
+ if (!job_assumes(j, runtime_add_mport(j->j_port, protocol_vproc_server, mxmsgsz) == KERN_SUCCESS)) {
+ goto out_bad2;
+ }
- if (!launchd_assumes(launchd_mport_notify_req(j->bs_port, MACH_NOTIFY_NO_SENDERS) == KERN_SUCCESS)) {
- launchd_assumes(launchd_mport_close_recv(j->bs_port) == KERN_SUCCESS);
+ if (!job_assumes(j, launchd_mport_notify_req(j->j_port, MACH_NOTIFY_NO_SENDERS) == KERN_SUCCESS)) {
+ (void)job_assumes(j, launchd_mport_close_recv(j->j_port) == KERN_SUCCESS);
goto out_bad;
}
- sprintf(j->label, "%d.%s", MACH_PORT_INDEX(j->bs_port), basename(j->argv[0]));
-
- job_log(j, LOG_INFO, "New%s server in bootstrap: %x", ond ? " on-demand" : "", jbs->bs_port);
-
- return j;
-
+ return true;
+out_bad2:
+ (void)job_assumes(j, launchd_mport_close_recv(j->j_port) == KERN_SUCCESS);
out_bad:
- if (j)
- job_remove(j);
- return NULL;
+ return false;
}
kern_return_t
-job_handle_mpm_wait(struct jobcb *j, mach_port_t srp, int *waitstatus)
+job_setup_exit_port(job_t j)
{
- if (j->p) {
- j->wait_reply_port = srp;
- return MIG_NO_REPLY;
+ kern_return_t kr = launchd_mport_create_recv(&j->exit_status_port);
+ if (!job_assumes(j, kr == KERN_SUCCESS)) {
+ return MACH_PORT_NULL;
}
- *waitstatus = j->last_exit_status;
+ struct mach_port_limits limits = {
+ .mpl_qlimit = 1,
+ };
+ kr = mach_port_set_attributes(mach_task_self(), j->exit_status_port, MACH_PORT_LIMITS_INFO, (mach_port_info_t)&limits, sizeof(limits));
+ (void)job_assumes(j, kr == KERN_SUCCESS);
- return 0;
+ kr = launchd_mport_make_send_once(j->exit_status_port, &j->exit_status_dest);
+ if (!job_assumes(j, kr == KERN_SUCCESS)) {
+ (void)job_assumes(j, launchd_mport_close_recv(j->exit_status_port) == KERN_SUCCESS);
+ j->exit_status_port = MACH_PORT_NULL;
+ }
+
+ return kr;
}
-struct jobcb *
-job_new_spawn(const char *label, const char *path, const char *workingdir, const char *const *argv, const char *const *env, mode_t *u_mask, bool w4d, bool fppc)
+job_t
+job_new_via_mach_init(job_t j, const char *cmd, uid_t uid, bool ond)
{
- struct jobcb *jr;
+ const char **argv = (const char **)mach_cmd2argv(cmd);
+ job_t jr = NULL;
- if ((jr = job_find(root_job, label)) != NULL) {
- errno = EEXIST;
- return NULL;
+ if (!job_assumes(j, argv != NULL)) {
+ goto out_bad;
}
- jr = job_new(root_job, label, path, argv, NULL, MACH_PORT_NULL);
+ jr = job_new(j->mgr, AUTO_PICK_LEGACY_LABEL, NULL, argv);
- if (!jr)
- return NULL;
-
- jr->unload_at_exit = true;
- jr->stall_before_exec = w4d;
- jr->force_ppc = fppc;
+ free(argv);
- if (!job_setup_machport(jr)) {
- job_remove(jr);
- return NULL;
+ /* jobs can easily be denied creation during shutdown */
+ if (unlikely(jr == NULL)) {
+ goto out_bad;
}
- if (workingdir)
- jr->workingdir = strdup(workingdir);
-
- if (u_mask) {
- jr->mask = *u_mask;
- jr->setmask = true;
- }
+ jr->mach_uid = uid;
+ jr->ondemand = ond;
+ jr->legacy_mach_job = true;
+ jr->abandon_pg = true;
+ jr->priv_port_has_senders = true; /* the IPC that called us will make-send on this port */
- if (env) for (; *env; env++) {
- char newkey[strlen(*env) + 1], *eqoff = strchr(*env, '=');
- if (!eqoff) {
- job_log(jr, LOG_WARNING, "Environmental variable missing '=' separator: %s", *env);
- continue;
- }
- strcpy(newkey, *env);
- *eqoff = '\0';
- envitem_new(jr, newkey, eqoff + 1, false);
+ if (!job_setup_machport(jr)) {
+ goto out_bad;
}
- job_start(jr);
+ job_log(jr, LOG_INFO, "Legacy%s server created", ond ? " on-demand" : "");
return jr;
+
+out_bad:
+ if (jr) {
+ job_remove(jr);
+ }
+ return NULL;
}
-struct jobcb *
-job_new(struct jobcb *p, const char *label, const char *prog, const char *const *argv, const char *stdinpath, mach_port_t reqport)
+job_t
+job_new_anonymous(jobmgr_t jm, pid_t anonpid)
{
- const char *const *argv_tmp = argv;
- char *co;
- int i, cc = 0;
- struct jobcb *j;
+ struct proc_bsdshortinfo proc;
+ bool shutdown_state;
+ job_t jp = NULL, jr = NULL;
+ uid_t kp_euid, kp_uid, kp_svuid;
+ gid_t kp_egid, kp_gid, kp_svgid;
- if (reqport == MACH_PORT_NULL && prog == NULL && argv == NULL) {
+ if (!jobmgr_assumes(jm, anonpid != 0)) {
errno = EINVAL;
return NULL;
}
-
- j = calloc(1, sizeof(struct jobcb) + strlen(label) + 1);
-
- if (!launchd_assumes(j != NULL))
- goto out_bad;
-
- strcpy(j->label, label);
- j->kqjob_callback = job_callback;
- j->parent = p ? job_get_bs(p) : NULL;
- j->ondemand = true;
- j->checkedin = true;
- j->firstborn = (strcmp(label, FIRSTBORN_LABEL) == 0);
-
- if (reqport != MACH_PORT_NULL) {
- j->req_port = reqport;
- if (!launchd_assumes(launchd_mport_notify_req(reqport, MACH_NOTIFY_DEAD_NAME) == KERN_SUCCESS))
- goto out_bad;
+
+ if (!jobmgr_assumes(jm, anonpid < 100000)) {
+ /* The kernel current defines PID_MAX to be 99999, but that define isn't exported */
+ errno = EINVAL;
+ return NULL;
}
- if (prog) {
- j->prog = strdup(prog);
- if (!launchd_assumes(j->prog != NULL))
- goto out_bad;
+ /* libproc returns the number of bytes written into the buffer upon success,
+ * zero on failure.
+ */
+ if (proc_pidinfo(anonpid, PROC_PIDT_SHORTBSDINFO, 1, &proc, PROC_PIDT_SHORTBSDINFO_SIZE) == 0) {
+ if (errno != ESRCH) {
+ (void)jobmgr_assumes(jm, errno == 0);
+ }
+ return NULL;
}
- if (stdinpath) {
- j->stdinpath = strdup(stdinpath);
- if (!launchd_assumes(j->stdinpath != NULL))
- goto out_bad;
+ if (!jobmgr_assumes(jm, proc.pbsi_comm[0] != '\0')) {
+ errno = EINVAL;
+ return NULL;
}
- if (argv) {
- while (*argv_tmp++)
- j->argc++;
-
- for (i = 0; i < j->argc; i++)
- cc += strlen(argv[i]) + 1;
-
- j->argv = malloc((j->argc + 1) * sizeof(char *) + cc);
+ if (unlikely(proc.pbsi_status == SZOMB)) {
+ jobmgr_log(jm, LOG_DEBUG, "Tried to create an anonymous job for zombie PID %u: %s", anonpid, proc.pbsi_comm);
+ }
- if (!launchd_assumes(j != NULL))
- goto out_bad;
+ if (unlikely(proc.pbsi_flags & P_SUGID)) {
+ jobmgr_log(jm, LOG_DEBUG, "Inconsistency: P_SUGID is set on PID %u: %s", anonpid, proc.pbsi_comm);
+ }
- co = ((char *)j->argv) + ((j->argc + 1) * sizeof(char *));
+ kp_euid = proc.pbsi_uid;
+ kp_uid = proc.pbsi_ruid;
+ kp_svuid = proc.pbsi_svuid;
+ kp_egid = proc.pbsi_gid;
+ kp_gid = proc.pbsi_rgid;
+ kp_svgid = proc.pbsi_svgid;
- for (i = 0; i < j->argc; i++) {
- j->argv[i] = co;
- strcpy(co, argv[i]);
- co += strlen(argv[i]) + 1;
- }
- j->argv[i] = NULL;
+ if (unlikely(kp_euid != kp_uid || kp_euid != kp_svuid || kp_uid != kp_svuid || kp_egid != kp_gid || kp_egid != kp_svgid || kp_gid != kp_svgid)) {
+ jobmgr_log(jm, LOG_DEBUG, "Inconsistency: Mixed credentials (e/r/s UID %u/%u/%u GID %u/%u/%u) detected on PID %u: %s",
+ kp_euid, kp_uid, kp_svuid, kp_egid, kp_gid, kp_svgid, anonpid, proc.pbsi_comm);
}
- if (j->parent) {
- SLIST_INSERT_HEAD(&j->parent->jobs, j, sle);
- job_log(j->parent, LOG_DEBUG, "Conceived");
+ /* "Fix" for a problem that shouldn't even exist.
+ * See rdar://problem/7264615 for the symptom and rdar://problem/5020256
+ * as to why this can happen.
+ */
+ if (!jobmgr_assumes(jm, (pid_t)proc.pbsi_ppid != anonpid)) {
+ jobmgr_log(jm, LOG_WARNING, "Process has become its own parent through ptrace(3). It should find a different way to do whatever it's doing. Setting PPID to 0: %s", proc.pbsi_comm);
+ errno = EINVAL;
+ return NULL;
}
- return j;
-
-out_bad:
- if (j) {
- if (j->prog)
- free(j->prog);
- if (j->stdinpath)
- free(j->stdinpath);
- free(j);
+ /* A total hack: Normally, job_new() returns an error during shutdown, but anonymous jobs are special. */
+ if (unlikely(shutdown_state = jm->shutting_down)) {
+ jm->shutting_down = false;
}
- return NULL;
-}
-struct jobcb *
-job_import(launch_data_t pload)
-{
- struct jobcb *j = job_import2(pload);
+ /* We only set requestor_pid for XPC domains. */
+ const char *whichlabel = (jm->req_pid == anonpid) ? AUTO_PICK_XPC_LABEL : AUTO_PICK_ANONYMOUS_LABEL;
+ if (jobmgr_assumes(jm, (jr = job_new(jm, whichlabel, proc.pbsi_comm, NULL)) != NULL)) {
+ u_int proc_fflags = NOTE_EXEC|NOTE_FORK|NOTE_EXIT;
- if (j == NULL)
- return NULL;
+ total_anon_children++;
+ jr->anonymous = true;
+ jr->p = anonpid;
- job_dispatch(j);
+ /* anonymous process reaping is messy */
+ LIST_INSERT_HEAD(&jm->active_jobs[ACTIVE_JOB_HASH(jr->p)], jr, pid_hash_sle);
- return j;
-}
+ if (unlikely(kevent_mod(jr->p, EVFILT_PROC, EV_ADD, proc_fflags, 0, root_jobmgr) == -1) && job_assumes(jr, errno == ESRCH)) {
+ /* zombies are weird */
+ job_log(jr, LOG_ERR, "Failed to add kevent for PID %u. Will unload at MIG return", jr->p);
+ jr->unload_at_mig_return = true;
+ }
-launch_data_t
-job_import_bulk(launch_data_t pload)
-{
- launch_data_t resp = launch_data_alloc(LAUNCH_DATA_ARRAY);
- struct jobcb **ja;
- size_t i, c = launch_data_array_get_count(pload);
+ if (unlikely(shutdown_state)) {
+ job_log(jr, LOG_SCOLDING, "This process showed up to the party while all the guests were leaving. Odds are that it will have a miserable time.");
+ }
- ja = alloca(c * sizeof(struct jobcb *));
+ job_log(jr, LOG_DEBUG, "Created PID %u anonymously by PPID %u%s%s", anonpid, proc.pbsi_ppid, jp ? ": " : "", jp ? jp->label : "");
+ }
- for (i = 0; i < c; i++) {
- if ((ja[i] = job_import2(launch_data_array_get_index(pload, i))))
- errno = 0;
- launch_data_array_set_index(resp, launch_data_new_errno(errno), i);
+ if (unlikely(shutdown_state)) {
+ jm->shutting_down = true;
}
- for (i = 0; i < c; i++) {
- if (ja[i] == NULL)
- continue;
- job_dispatch(ja[i]);
+ /* This is down here to mitigate the effects of rdar://problem/7264615, in which a process
+ * attaches to its own parent. We need to make sure that the anonymous job has been added
+ * to the process list so that, if it's used ptrace(3) to cause a cycle in the process
+ * tree (thereby making it not a tree anymore), we'll find the tracing parent PID of the
+ * parent process, which is the child, when we go looking for it in jobmgr_find_by_pid().
+ */
+ switch (proc.pbsi_ppid) {
+ case 0:
+ /* the kernel */
+ break;
+ case 1:
+ if (!pid1_magic) {
+ /* we cannot possibly find a parent job_t that is useful in this function */
+ break;
+ }
+ /* fall through */
+ default:
+ jp = jobmgr_find_by_pid(jm, proc.pbsi_ppid, true);
+ if (jobmgr_assumes(jm, jp != NULL)) {
+ if (jp && !jp->anonymous && unlikely(!(proc.pbsi_flags & P_EXEC))) {
+ job_log(jp, LOG_DEBUG, "Called *fork(). Please switch to posix_spawn*(), pthreads or launchd. Child PID %u", proc.pbsi_pid);
+ }
+ }
+ break;
}
- return resp;
+ return jr;
}
-void
-job_import_bool(struct jobcb *j, const char *key, bool value)
+job_t
+job_new_subjob(job_t j, uuid_t identifier)
{
- switch (key[0]) {
- case 'f':
- case 'F':
- if (strcasecmp(key, LAUNCH_JOBKEY_FORCEPOWERPC) == 0)
- j->force_ppc = value;
- break;
- case 'k':
- case 'K':
- if (strcasecmp(key, LAUNCH_JOBKEY_KEEPALIVE) == 0)
- j->ondemand = !value;
- break;
- case 'o':
- case 'O':
- if (strcasecmp(key, LAUNCH_JOBKEY_ONDEMAND) == 0)
+ char label[0];
+ uuid_string_t idstr;
+ uuid_unparse(identifier, idstr);
+ size_t label_sz = snprintf(label, 0, "%s.%s", j->label, idstr);
+
+ job_t nj = (struct job_s *)calloc(1, sizeof(struct job_s) + label_sz + 1);
+ if (launchd_assumes(nj != NULL)) {
+ nj->kqjob_callback = job_callback;
+ nj->mgr = j->mgr;
+ nj->min_run_time = j->min_run_time;
+ nj->timeout = j->timeout;
+ nj->exit_timeout = j->exit_timeout;
+
+ snprintf((char *)nj->label, label_sz + 1, "%s.%s", j->label, idstr);
+
+ /* Set all our simple Booleans that are applicable. */
+ nj->debug = j->debug;
+ nj->ondemand = j->ondemand;
+ nj->checkedin = true;
+ nj->low_pri_io = j->low_pri_io;
+ nj->setmask = j->setmask;
+ nj->wait4debugger = j->wait4debugger;
+ nj->internal_exc_handler = j->internal_exc_handler;
+ nj->setnice = j->setnice;
+ nj->abandon_pg = j->abandon_pg;
+ nj->ignore_pg_at_shutdown = j->ignore_pg_at_shutdown;
+ nj->deny_job_creation = j->deny_job_creation;
+ nj->kill_via_shmem = j->kill_via_shmem;
+ nj->needs_kickoff = j->needs_kickoff;
+ nj->currently_ignored = true;
+ nj->dedicated_instance = true;
+ nj->xpc_service = j->xpc_service;
+
+ nj->mask = j->mask;
+ uuid_copy(nj->instance_id, identifier);
+
+ /* These jobs are purely on-demand Mach jobs. */
+
+ /* {Hard | Soft}ResourceLimits are not supported. */
+
+ struct machservice *msi = NULL;
+ SLIST_FOREACH(msi, &j->machservices, sle) {
+ /* Only copy MachServices that were actually declared in the plist.
+ * So skip over per-PID ones and ones that were created via
+ * bootstrap_register().
+ */
+ if (msi->upfront) {
+ mach_port_t mp = MACH_PORT_NULL;
+ struct machservice *msj = machservice_new(nj, msi->name, &mp, msi->per_pid);
+ if (job_assumes(nj, msj != NULL)) {
+ msj->reset = msi->reset;
+ msj->delete_on_destruction = msi->delete_on_destruction;
+ msj->drain_one_on_crash = msi->drain_one_on_crash;
+ msj->drain_all_on_crash = msi->drain_all_on_crash;
+ }
+ }
+ }
+
+ if (j->prog) {
+ nj->prog = strdup(j->prog);
+ }
+ if (j->argv) {
+ size_t sz = malloc_size(j->argv);
+ nj->argv = (char **)malloc(sz);
+ if (job_assumes(nj, nj->argv != NULL)) {
+ /* This is the start of our strings. */
+ char *p = ((char *)nj->argv) + ((j->argc + 1) * sizeof(char *));
+
+ size_t i = 0;
+ for (i = 0; i < j->argc; i++) {
+ (void)strcpy(p, j->argv[i]);
+ nj->argv[i] = p;
+ p += (strlen(j->argv[i]) + 1);
+ }
+ nj->argv[i] = NULL;
+ }
+
+ nj->argc = j->argc;
+ }
+
+ /* We ignore global environment variables. */
+ struct envitem *ei = NULL;
+ SLIST_FOREACH(ei, &j->env, sle) {
+ (void)job_assumes(nj, envitem_new(nj, ei->key, ei->value, false, false));
+ }
+ uuid_string_t val;
+ uuid_unparse(identifier, val);
+ (void)job_assumes(nj, envitem_new(nj, LAUNCH_ENV_INSTANCEID, val, false, false));
+
+ if (j->rootdir) {
+ nj->rootdir = strdup(j->rootdir);
+ }
+ if (j->workingdir) {
+ nj->workingdir = strdup(j->workingdir);
+ }
+ if (j->username) {
+ nj->username = strdup(j->username);
+ }
+ if (j->groupname) {
+ nj->groupname = strdup(j->groupname);
+ }
+ /* FIXME: We shouldn't redirect all the output from these jobs to the same
+ * file. We should uniquify the file names.
+ */
+ if (j->stdinpath) {
+ nj->stdinpath = strdup(j->stdinpath);
+ }
+ if (j->stdoutpath) {
+ nj->stdoutpath = strdup(j->stdinpath);
+ }
+ if (j->stderrpath) {
+ nj->stderrpath = strdup(j->stderrpath);
+ }
+ if (j->alt_exc_handler) {
+ nj->alt_exc_handler = strdup(j->alt_exc_handler);
+ }
+ #if HAVE_SANDBOX
+ if (j->seatbelt_profile) {
+ nj->seatbelt_profile = strdup(j->seatbelt_profile);
+ }
+ #endif
+
+ #if HAVE_QUARANTINE
+ if (j->quarantine_data) {
+ nj->quarantine_data = strdup(j->quarantine_data);
+ }
+ nj->quarantine_data_sz = j->quarantine_data_sz;
+ #endif
+ if (j->j_binpref) {
+ size_t sz = malloc_size(j->j_binpref);
+ nj->j_binpref = (cpu_type_t *)malloc(sz);
+ if (job_assumes(nj, nj->j_binpref)) {
+ memcpy(&nj->j_binpref, &j->j_binpref, sz);
+ }
+ }
+
+ /* JetsamPriority is unsupported. */
+
+ if (j->asport != MACH_PORT_NULL) {
+ (void)job_assumes(nj, launchd_mport_copy_send(j->asport) == KERN_SUCCESS);
+ nj->asport = j->asport;
+ }
+
+ LIST_INSERT_HEAD(&nj->mgr->jobs, nj, sle);
+
+ jobmgr_t where2put = root_jobmgr;
+ if (j->mgr->properties & BOOTSTRAP_PROPERTY_XPC_DOMAIN) {
+ where2put = j->mgr;
+ }
+ LIST_INSERT_HEAD(&where2put->label_hash[hash_label(nj->label)], nj, label_hash_sle);
+ LIST_INSERT_HEAD(&j->subjobs, nj, subjob_sle);
+ }
+
+ return nj;
+}
+
+job_t
+job_new(jobmgr_t jm, const char *label, const char *prog, const char *const *argv)
+{
+ const char *const *argv_tmp = argv;
+ char tmp_path[PATH_MAX];
+ char auto_label[1000];
+ const char *bn = NULL;
+ char *co;
+ size_t minlabel_len;
+ size_t i, cc = 0;
+ job_t j;
+
+ launchd_assert(offsetof(struct job_s, kqjob_callback) == 0);
+
+ if (unlikely(jm->shutting_down)) {
+ errno = EINVAL;
+ return NULL;
+ }
+
+ if (unlikely(prog == NULL && argv == NULL)) {
+ errno = EINVAL;
+ return NULL;
+ }
+
+ char *anon_or_legacy = (label == AUTO_PICK_ANONYMOUS_LABEL) ? "anonymous" : "mach_init";
+ if (unlikely(label == AUTO_PICK_LEGACY_LABEL || label == AUTO_PICK_ANONYMOUS_LABEL)) {
+ if (prog) {
+ bn = prog;
+ } else {
+ strlcpy(tmp_path, argv[0], sizeof(tmp_path));
+ bn = basename(tmp_path); /* prog for auto labels is kp.kp_kproc.p_comm */
+ }
+ snprintf(auto_label, sizeof(auto_label), "%s.%s.%s", sizeof(void *) == 8 ? "0xdeadbeeffeedface" : "0xbabecafe", anon_or_legacy, bn);
+ label = auto_label;
+ /* This is so we can do gross things later. See NOTE_EXEC for anonymous jobs */
+ minlabel_len = strlen(label) + MAXCOMLEN;
+ } else {
+ if (label == AUTO_PICK_XPC_LABEL) {
+ minlabel_len = snprintf(auto_label, sizeof(auto_label), "com.apple.xpc.domain-owner.%s", jm->owner);
+ } else {
+ minlabel_len = strlen(label);
+ }
+ }
+
+ j = calloc(1, sizeof(struct job_s) + minlabel_len + 1);
+
+ if (!jobmgr_assumes(jm, j != NULL)) {
+ return NULL;
+ }
+
+ if (unlikely(label == auto_label)) {
+ snprintf((char *)j->label, strlen(label) + 1, "%p.%s.%s", j, anon_or_legacy, bn);
+ } else {
+ strcpy((char *)j->label, (label == AUTO_PICK_XPC_LABEL) ? auto_label : label);
+ }
+ j->kqjob_callback = job_callback;
+ j->mgr = jm;
+ j->min_run_time = LAUNCHD_MIN_JOB_RUN_TIME;
+ j->timeout = RUNTIME_ADVISABLE_IDLE_TIMEOUT;
+ j->exit_timeout = LAUNCHD_DEFAULT_EXIT_TIMEOUT;
+ j->currently_ignored = true;
+ j->ondemand = true;
+ j->checkedin = true;
+ j->jetsam_priority = -1;
+ j->jetsam_memlimit = -1;
+ j->jetsam_seq = -1;
+ uuid_clear(j->expected_audit_uuid);
+
+ if (prog) {
+ j->prog = strdup(prog);
+ if (!job_assumes(j, j->prog != NULL)) {
+ goto out_bad;
+ }
+ }
+
+ if (likely(argv)) {
+ while (*argv_tmp++) {
+ j->argc++;
+ }
+
+ for (i = 0; i < j->argc; i++) {
+ cc += strlen(argv[i]) + 1;
+ }
+
+ j->argv = malloc((j->argc + 1) * sizeof(char *) + cc);
+
+ if (!job_assumes(j, j->argv != NULL)) {
+ goto out_bad;
+ }
+
+ co = ((char *)j->argv) + ((j->argc + 1) * sizeof(char *));
+
+ for (i = 0; i < j->argc; i++) {
+ j->argv[i] = co;
+ strcpy(co, argv[i]);
+ co += strlen(argv[i]) + 1;
+ }
+ j->argv[i] = NULL;
+ }
+
+ if (strcmp(j->label, "com.apple.WindowServer") == 0) {
+ j->has_console = true;
+ }
+
+ LIST_INSERT_HEAD(&jm->jobs, j, sle);
+
+ jobmgr_t where2put_label = root_jobmgr;
+ if (j->mgr->properties & BOOTSTRAP_PROPERTY_XPC_DOMAIN) {
+ where2put_label = j->mgr;
+ }
+ LIST_INSERT_HEAD(&where2put_label->label_hash[hash_label(j->label)], j, label_hash_sle);
+ uuid_clear(j->expected_audit_uuid);
+
+ job_log(j, LOG_DEBUG, "Conceived");
+
+ return j;
+
+out_bad:
+ if (j->prog) {
+ free(j->prog);
+ }
+ free(j);
+
+ return NULL;
+}
+
+#ifndef __LAUNCH_DISABLE_XPC_SUPPORT__
+job_t
+job_new_alias(jobmgr_t jm, job_t src)
+{
+ job_t j = NULL;
+ if (job_find(jm, src->label)) {
+ errno = EEXIST;
+ } else {
+ j = calloc(1, sizeof(struct job_s) + strlen(src->label) + 1);
+ if (jobmgr_assumes(jm, j != NULL)) {
+ strcpy((char *)j->label, src->label);
+ LIST_INSERT_HEAD(&jm->jobs, j, sle);
+ LIST_INSERT_HEAD(&jm->label_hash[hash_label(j->label)], j, label_hash_sle);
+ /* Bad jump address. The kqueue callback for aliases should never be
+ * invoked.
+ */
+ j->kqjob_callback = (kq_callback)0xfa1afe1;
+ j->alias = src;
+ j->mgr = jm;
+
+ struct machservice *msi = NULL;
+ SLIST_FOREACH(msi, &src->machservices, sle) {
+ if (!machservice_new_alias(j, msi)) {
+ jobmgr_log(jm, LOG_ERR, "Failed to alias job: %s", src->label);
+ errno = EINVAL;
+ job_remove(j);
+ j = NULL;
+ break;
+ }
+ }
+ }
+
+ if (j) {
+ job_log(j, LOG_DEBUG, "Aliased service into domain: %s", jm->name);
+ }
+ }
+
+ return j;
+}
+#endif
+
+job_t
+job_import(launch_data_t pload)
+{
+ job_t j = jobmgr_import2(root_jobmgr, pload);
+
+ if (unlikely(j == NULL)) {
+ return NULL;
+ }
+
+ /* Since jobs are effectively stalled until they get security sessions assigned
+ * to them, we may wish to reconsider this behavior of calling the job "enabled"
+ * as far as other jobs with the OtherJobEnabled KeepAlive criterion set.
+ */
+ job_dispatch_curious_jobs(j);
+ return job_dispatch(j, false);
+}
+
+launch_data_t
+job_import_bulk(launch_data_t pload)
+{
+ launch_data_t resp = launch_data_alloc(LAUNCH_DATA_ARRAY);
+ job_t *ja;
+ size_t i, c = launch_data_array_get_count(pload);
+
+ ja = alloca(c * sizeof(job_t));
+
+ for (i = 0; i < c; i++) {
+ if ((likely(ja[i] = jobmgr_import2(root_jobmgr, launch_data_array_get_index(pload, i)))) && errno != ENEEDAUTH) {
+ errno = 0;
+ }
+ launch_data_array_set_index(resp, launch_data_new_errno(errno), i);
+ }
+
+ for (i = 0; i < c; i++) {
+ if (likely(ja[i])) {
+ job_dispatch_curious_jobs(ja[i]);
+ job_dispatch(ja[i], false);
+ }
+ }
+
+ return resp;
+}
+
+void
+job_import_bool(job_t j, const char *key, bool value)
+{
+ bool found_key = false;
+
+ switch (key[0]) {
+ case 'a':
+ case 'A':
+ if (strcasecmp(key, LAUNCH_JOBKEY_ABANDONPROCESSGROUP) == 0) {
+ j->abandon_pg = value;
+ found_key = true;
+ }
+ break;
+ case 'b':
+ case 'B':
+ if (strcasecmp(key, LAUNCH_JOBKEY_BEGINTRANSACTIONATSHUTDOWN) == 0) {
+ j->dirty_at_shutdown = value;
+ found_key = true;
+ }
+ break;
+ case 'k':
+ case 'K':
+ if (strcasecmp(key, LAUNCH_JOBKEY_KEEPALIVE) == 0) {
+ j->ondemand = !value;
+ found_key = true;
+ }
+ break;
+ case 'o':
+ case 'O':
+ if (strcasecmp(key, LAUNCH_JOBKEY_ONDEMAND) == 0) {
j->ondemand = value;
+ found_key = true;
+ }
break;
case 'd':
case 'D':
- if (strcasecmp(key, LAUNCH_JOBKEY_DEBUG) == 0)
+ if (strcasecmp(key, LAUNCH_JOBKEY_DEBUG) == 0) {
j->debug = value;
+ found_key = true;
+ } else if (strcasecmp(key, LAUNCH_JOBKEY_DISABLED) == 0) {
+ (void)job_assumes(j, !value);
+ found_key = true;
+ } else if (strcasecmp(key, LAUNCH_JOBKEY_DISABLEASLR) == 0) {
+ j->disable_aslr = value;
+ found_key = true;
+ }
+ break;
+ case 'h':
+ case 'H':
+ if (strcasecmp(key, LAUNCH_JOBKEY_HOPEFULLYEXITSLAST) == 0) {
+ job_log(j, LOG_INFO, "%s has been deprecated. Please use the new %s key instead and add EnableTransactions to your launchd.plist.", LAUNCH_JOBKEY_HOPEFULLYEXITSLAST, LAUNCH_JOBKEY_BEGINTRANSACTIONATSHUTDOWN);
+ j->dirty_at_shutdown = value;
+ found_key = true;
+ }
break;
case 's':
case 'S':
- if (strcasecmp(key, LAUNCH_JOBKEY_SESSIONCREATE) == 0)
+ if (strcasecmp(key, LAUNCH_JOBKEY_SESSIONCREATE) == 0) {
j->session_create = value;
+ found_key = true;
+ } else if (strcasecmp(key, LAUNCH_JOBKEY_STARTONMOUNT) == 0) {
+ j->start_on_mount = value;
+ found_key = true;
+ } else if (strcasecmp(key, LAUNCH_JOBKEY_SERVICEIPC) == 0) {
+ /* this only does something on Mac OS X 10.4 "Tiger" */
+ found_key = true;
+ } else if (strcasecmp(key, LAUNCH_JOBKEY_SHUTDOWNMONITOR) == 0) {
+ if (_s_shutdown_monitor) {
+ job_log(j, LOG_ERR, "Only one job may monitor shutdown.");
+ } else {
+ j->shutdown_monitor = true;
+ _s_shutdown_monitor = j;
+ }
+ found_key = true;
+ }
break;
case 'l':
case 'L':
- if (strcasecmp(key, LAUNCH_JOBKEY_LOWPRIORITYIO) == 0)
+ if (strcasecmp(key, LAUNCH_JOBKEY_LOWPRIORITYIO) == 0) {
j->low_pri_io = value;
+ found_key = true;
+ } else if (strcasecmp(key, LAUNCH_JOBKEY_LAUNCHONLYONCE) == 0) {
+ j->only_once = value;
+ found_key = true;
+ }
+ break;
+ case 'm':
+ case 'M':
+ if (strcasecmp(key, LAUNCH_JOBKEY_MACHEXCEPTIONHANDLER) == 0) {
+ j->internal_exc_handler = value;
+ found_key = true;
+ } else if (strcasecmp(key, LAUNCH_JOBKEY_MULTIPLEINSTANCES) == 0) {
+ j->multiple_instances = value;
+ found_key = true;
+ }
break;
case 'i':
case 'I':
- if (strcasecmp(key, LAUNCH_JOBKEY_INITGROUPS) == 0)
- j->init_groups = value;
+ if (strcasecmp(key, LAUNCH_JOBKEY_INITGROUPS) == 0) {
+ if (getuid() != 0) {
+ job_log(j, LOG_WARNING, "Ignored this key: %s", key);
+ return;
+ }
+ j->no_init_groups = !value;
+ found_key = true;
+ } else if (strcasecmp(key, LAUNCH_JOBKEY_IGNOREPROCESSGROUPATSHUTDOWN) == 0) {
+ j->ignore_pg_at_shutdown = value;
+ found_key = true;
+ }
break;
case 'r':
case 'R':
- if (strcasecmp(key, LAUNCH_JOBKEY_RUNATLOAD) == 0)
- j->runatload = value;
+ if (strcasecmp(key, LAUNCH_JOBKEY_RUNATLOAD) == 0) {
+ if (value) {
+ /* We don't want value == false to change j->start_pending */
+ j->start_pending = true;
+ }
+ found_key = true;
+ }
break;
case 'e':
case 'E':
- if (strcasecmp(key, LAUNCH_JOBKEY_ENABLEGLOBBING) == 0)
+ if (strcasecmp(key, LAUNCH_JOBKEY_ENABLEGLOBBING) == 0) {
j->globargv = value;
+ found_key = true;
+ } else if (strcasecmp(key, LAUNCH_JOBKEY_ENABLETRANSACTIONS) == 0) {
+ j->kill_via_shmem = value;
+ found_key = true;
+ } else if (strcasecmp(key, LAUNCH_JOBKEY_ENTERKERNELDEBUGGERBEFOREKILL) == 0) {
+ j->debug_before_kill = value;
+ found_key = true;
+ } else if (strcasecmp(key, LAUNCH_JOBKEY_EMBEDDEDPRIVILEGEDISPENSATION) == 0) {
+ if (!s_embedded_privileged_job) {
+ j->embedded_special_privileges = value;
+ s_embedded_privileged_job = j;
+ } else {
+ job_log(j, LOG_ERR, "Job tried to claim %s after it has already been claimed.", key);
+ }
+ found_key = true;
+ } else if (strcasecmp(key, LAUNCH_JOBKEY_EVENTMONITOR) == 0) {
+ if (job_assumes(j, _s_event_monitor == NULL)) {
+ j->event_monitor = value;
+ if (value) {
+ _s_event_monitor = j;
+ }
+ } else {
+ job_log(j, LOG_NOTICE, "Job tried to steal event monitoring responsibility!");
+ }
+ found_key = true;
+ }
break;
case 'w':
case 'W':
- if (strcasecmp(key, LAUNCH_JOBKEY_WAITFORDEBUGGER) == 0)
+ if (strcasecmp(key, LAUNCH_JOBKEY_WAITFORDEBUGGER) == 0) {
j->wait4debugger = value;
+ found_key = true;
+ }
break;
default:
break;
}
+
+ if (unlikely(!found_key)) {
+ job_log(j, LOG_WARNING, "Unknown key for boolean: %s", key);
+ }
}
void
-job_import_string(struct jobcb *j, const char *key, const char *value)
+job_import_string(job_t j, const char *key, const char *value)
{
char **where2put = NULL;
- char **ignore = (char **)-1;
switch (key[0]) {
+ case 'm':
+ case 'M':
+ if (strcasecmp(key, LAUNCH_JOBKEY_MACHEXCEPTIONHANDLER) == 0) {
+ where2put = &j->alt_exc_handler;
+ }
+ break;
case 'p':
case 'P':
- if (strcasecmp(key, LAUNCH_JOBKEY_PROGRAM) == 0)
- where2put = ignore;
+ if (strcasecmp(key, LAUNCH_JOBKEY_PROGRAM) == 0) {
+ return;
+ } else if (strcasecmp(key, LAUNCH_JOBKEY_POSIXSPAWNTYPE) == 0) {
+ if (strcasecmp(value, LAUNCH_KEY_POSIXSPAWNTYPE_TALAPP) == 0) {
+ j->pstype = POSIX_SPAWN_OSX_TALAPP_START;
+ } else if (strcasecmp(value, LAUNCH_KEY_POSIXSPAWNTYPE_WIDGET) == 0) {
+ j->pstype = POSIX_SPAWN_OSX_WIDGET_START;
+ }
+#if TARGET_OS_EMBEDDED
+ else if (strcasecmp(value, LAUNCH_KEY_POSIXSPAWNTYPE_IOSAPP) == 0) {
+ j->pstype = POSIX_SPAWN_IOS_APP_START;
+ }
+#endif
+ else {
+ job_log(j, LOG_ERR, "Unknown value for key %s: %s", key, value);
+ }
+ return;
+ }
break;
case 'l':
case 'L':
- if (strcasecmp(key, LAUNCH_JOBKEY_LABEL) == 0)
- where2put = ignore;
+ if (strcasecmp(key, LAUNCH_JOBKEY_LABEL) == 0) {
+ return;
+ } else if (strcasecmp(key, LAUNCH_JOBKEY_LIMITLOADTOHOSTS) == 0) {
+ return;
+ } else if (strcasecmp(key, LAUNCH_JOBKEY_LIMITLOADFROMHOSTS) == 0) {
+ return;
+ } else if (strcasecmp(key, LAUNCH_JOBKEY_LIMITLOADTOSESSIONTYPE) == 0) {
+ return;
+ }
break;
case 'r':
case 'R':
- if (strcasecmp(key, LAUNCH_JOBKEY_ROOTDIRECTORY) == 0)
+ if (strcasecmp(key, LAUNCH_JOBKEY_ROOTDIRECTORY) == 0) {
+ if (getuid() != 0) {
+ job_log(j, LOG_WARNING, "Ignored this key: %s", key);
+ return;
+ }
where2put = &j->rootdir;
+ }
break;
case 'w':
case 'W':
- if (strcasecmp(key, LAUNCH_JOBKEY_WORKINGDIRECTORY) == 0)
+ if (strcasecmp(key, LAUNCH_JOBKEY_WORKINGDIRECTORY) == 0) {
where2put = &j->workingdir;
+ }
break;
case 'u':
case 'U':
- if (strcasecmp(key, LAUNCH_JOBKEY_USERNAME) == 0)
+ if (strcasecmp(key, LAUNCH_JOBKEY_USERNAME) == 0) {
+ if (getuid() != 0) {
+ job_log(j, LOG_WARNING, "Ignored this key: %s", key);
+ return;
+ } else if (strcmp(value, "root") == 0) {
+ return;
+ }
where2put = &j->username;
+ }
break;
case 'g':
case 'G':
- if (strcasecmp(key, LAUNCH_JOBKEY_GROUPNAME) == 0)
+ if (strcasecmp(key, LAUNCH_JOBKEY_GROUPNAME) == 0) {
+ if (getuid() != 0) {
+ job_log(j, LOG_WARNING, "Ignored this key: %s", key);
+ return;
+ } else if (strcmp(value, "wheel") == 0) {
+ return;
+ }
where2put = &j->groupname;
+ }
break;
case 's':
case 'S':
where2put = &j->stdoutpath;
} else if (strcasecmp(key, LAUNCH_JOBKEY_STANDARDERRORPATH) == 0) {
where2put = &j->stderrpath;
+ } else if (strcasecmp(key, LAUNCH_JOBKEY_STANDARDINPATH) == 0) {
+ where2put = &j->stdinpath;
+ j->stdin_fd = _fd(open(value, O_RDONLY|O_CREAT|O_NOCTTY|O_NONBLOCK, DEFFILEMODE));
+ if (job_assumes(j, j->stdin_fd != -1)) {
+ /* open() should not block, but regular IO by the job should */
+ (void)job_assumes(j, fcntl(j->stdin_fd, F_SETFL, 0) != -1);
+ /* XXX -- EV_CLEAR should make named pipes happy? */
+ (void)job_assumes(j, kevent_mod(j->stdin_fd, EVFILT_READ, EV_ADD|EV_CLEAR, 0, 0, j) != -1);
+ } else {
+ j->stdin_fd = 0;
+ }
+#if HAVE_SANDBOX
+ } else if (strcasecmp(key, LAUNCH_JOBKEY_SANDBOXPROFILE) == 0) {
+ where2put = &j->seatbelt_profile;
+#endif
+ }
+ break;
+ case 'X':
+ case 'x':
+ if (strcasecmp(key, LAUNCH_JOBKEY_XPCDOMAIN) == 0) {
+ return;
}
break;
default:
+ job_log(j, LOG_WARNING, "Unknown key for string: %s", key);
break;
}
- if (where2put) {
- if (where2put == ignore)
- return;
-
- launchd_assumes((*where2put = strdup(value)) != NULL);
+ if (likely(where2put)) {
+ (void)job_assumes(j, (*where2put = strdup(value)) != NULL);
} else {
- job_log(j, LOG_WARNING, "Unknown value for key %s: %s", key, value);
+ /* See rdar://problem/5496612. These two are okay. */
+ if (strncmp(key, "SHAuthorizationRight", sizeof("SHAuthorizationRight")) != 0 && strncmp(key, "ServiceDescription", sizeof("ServiceDescription")) != 0) {
+ job_log(j, LOG_WARNING, "Unknown key: %s", key);
+ }
}
}
void
-job_import_integer(struct jobcb *j, const char *key, long long value)
+job_import_integer(job_t j, const char *key, long long value)
{
switch (key[0]) {
+ case 'e':
+ case 'E':
+ if (strcasecmp(key, LAUNCH_JOBKEY_EXITTIMEOUT) == 0) {
+ if (unlikely(value < 0)) {
+ job_log(j, LOG_WARNING, "%s less than zero. Ignoring.", LAUNCH_JOBKEY_EXITTIMEOUT);
+ } else if (unlikely(value > UINT32_MAX)) {
+ job_log(j, LOG_WARNING, "%s is too large. Ignoring.", LAUNCH_JOBKEY_EXITTIMEOUT);
+ } else {
+ j->exit_timeout = (typeof(j->exit_timeout)) value;
+ }
+ } else if (strcasecmp(key, LAUNCH_JOBKEY_EMBEDDEDMAINTHREADPRIORITY) == 0) {
+ j->main_thread_priority = value;
+ }
+ break;
+ case 'j':
+ case 'J':
+ if (strcasecmp(key, LAUNCH_JOBKEY_JETSAMPRIORITY) == 0) {
+ job_log(j, LOG_WARNING | LOG_CONSOLE, "Please change the JetsamPriority key to be in a dictionary named JetsamProperties.");
+
+ launch_data_t pri = launch_data_new_integer(value);
+ if (job_assumes(j, pri != NULL)) {
+ jetsam_property_setup(pri, LAUNCH_JOBKEY_JETSAMPRIORITY, j);
+ launch_data_free(pri);
+ }
+ }
case 'n':
case 'N':
- if (strcasecmp(key, LAUNCH_JOBKEY_NICE) == 0)
- j->nice = value;
+ if (strcasecmp(key, LAUNCH_JOBKEY_NICE) == 0) {
+ if (unlikely(value < PRIO_MIN)) {
+ job_log(j, LOG_WARNING, "%s less than %d. Ignoring.", LAUNCH_JOBKEY_NICE, PRIO_MIN);
+ } else if (unlikely(value > PRIO_MAX)) {
+ job_log(j, LOG_WARNING, "%s is greater than %d. Ignoring.", LAUNCH_JOBKEY_NICE, PRIO_MAX);
+ } else {
+ j->nice = (typeof(j->nice)) value;
+ j->setnice = true;
+ }
+ }
break;
case 't':
case 'T':
if (strcasecmp(key, LAUNCH_JOBKEY_TIMEOUT) == 0) {
- if (value <= 0)
- job_log(j, LOG_WARNING, "Timeout less than or equal to zero. Ignoring.");
- else
- j->timeout = value;
+ if (unlikely(value < 0)) {
+ job_log(j, LOG_WARNING, "%s less than zero. Ignoring.", LAUNCH_JOBKEY_TIMEOUT);
+ } else if (unlikely(value > UINT32_MAX)) {
+ job_log(j, LOG_WARNING, "%s is too large. Ignoring.", LAUNCH_JOBKEY_TIMEOUT);
+ } else {
+ j->timeout = (typeof(j->timeout)) value;
+ }
+ } else if (strcasecmp(key, LAUNCH_JOBKEY_THROTTLEINTERVAL) == 0) {
+ if (value < 0) {
+ job_log(j, LOG_WARNING, "%s less than zero. Ignoring.", LAUNCH_JOBKEY_THROTTLEINTERVAL);
+ } else if (value > UINT32_MAX) {
+ job_log(j, LOG_WARNING, "%s is too large. Ignoring.", LAUNCH_JOBKEY_THROTTLEINTERVAL);
+ } else {
+ j->min_run_time = (typeof(j->min_run_time)) value;
+ }
}
break;
case 'u':
case 's':
case 'S':
if (strcasecmp(key, LAUNCH_JOBKEY_STARTINTERVAL) == 0) {
- if (value <= 0)
- job_log(j, LOG_WARNING, "StartInterval is not greater than zero, ignoring");
- else
- j->start_interval = value;
- if (-1 == kevent_mod((uintptr_t)&j->start_interval, EVFILT_TIMER, EV_ADD, NOTE_SECONDS, value, j))
- job_log_error(j, LOG_ERR, "adding kevent timer");
+ if (unlikely(value <= 0)) {
+ job_log(j, LOG_WARNING, "%s is not greater than zero. Ignoring.", LAUNCH_JOBKEY_STARTINTERVAL);
+ } else if (unlikely(value > UINT32_MAX)) {
+ job_log(j, LOG_WARNING, "%s is too large. Ignoring.", LAUNCH_JOBKEY_STARTINTERVAL);
+ } else {
+ runtime_add_weak_ref();
+ j->start_interval = (typeof(j->start_interval)) value;
+
+ (void)job_assumes(j, kevent_mod((uintptr_t)&j->start_interval, EVFILT_TIMER, EV_ADD, NOTE_SECONDS, j->start_interval, j) != -1);
+ }
+#if HAVE_SANDBOX
+ } else if (strcasecmp(key, LAUNCH_JOBKEY_SANDBOXFLAGS) == 0) {
+ j->seatbelt_flags = value;
+#endif
}
+
break;
default:
+ job_log(j, LOG_WARNING, "Unknown key for integer: %s", key);
break;
}
}
void
-job_import_dictionary(struct jobcb *j, const char *key, launch_data_t value)
+job_import_opaque(job_t j __attribute__((unused)),
+ const char *key, launch_data_t value __attribute__((unused)))
{
- launch_data_t tmp;
-
switch (key[0]) {
- case 'k':
- case 'K':
- if (strcasecmp(key, LAUNCH_JOBKEY_KEEPALIVE) == 0)
- launch_data_dict_iterate(value, semaphoreitem_setup, j);
- break;
- case 'i':
- case 'I':
- if (strcasecmp(key, LAUNCH_JOBKEY_INETDCOMPATIBILITY) == 0) {
- j->inetcompat = true;
- if ((tmp = launch_data_dict_lookup(value, LAUNCH_JOBINETDCOMPATIBILITY_WAIT)))
- j->inetcompat_wait = launch_data_get_bool(tmp);
- }
- break;
- case 'e':
+ case 'q':
+ case 'Q':
+#if HAVE_QUARANTINE
+ if (strcasecmp(key, LAUNCH_JOBKEY_QUARANTINEDATA) == 0) {
+ size_t tmpsz = launch_data_get_opaque_size(value);
+
+ if (job_assumes(j, j->quarantine_data = malloc(tmpsz))) {
+ memcpy(j->quarantine_data, launch_data_get_opaque(value), tmpsz);
+ j->quarantine_data_sz = tmpsz;
+ }
+ }
+#endif
+ case 's':
+ case 'S':
+ if (strcasecmp(key, LAUNCH_JOBKEY_SECURITYSESSIONUUID) == 0) {
+ size_t tmpsz = launch_data_get_opaque_size(value);
+ if (job_assumes(j, tmpsz == sizeof(uuid_t))) {
+ memcpy(j->expected_audit_uuid, launch_data_get_opaque(value), sizeof(uuid_t));
+ }
+ }
+ break;
+ default:
+ break;
+ }
+}
+
+static void
+policy_setup(launch_data_t obj, const char *key, void *context)
+{
+ job_t j = context;
+ bool found_key = false;
+
+ switch (key[0]) {
+ case 'd':
+ case 'D':
+ if (strcasecmp(key, LAUNCH_JOBPOLICY_DENYCREATINGOTHERJOBS) == 0) {
+ j->deny_job_creation = launch_data_get_bool(obj);
+ found_key = true;
+ }
+ break;
+ default:
+ break;
+ }
+
+ if (unlikely(!found_key)) {
+ job_log(j, LOG_WARNING, "Unknown policy: %s", key);
+ }
+}
+
+void
+job_import_dictionary(job_t j, const char *key, launch_data_t value)
+{
+ launch_data_t tmp;
+
+ switch (key[0]) {
+ case 'p':
+ case 'P':
+ if (strcasecmp(key, LAUNCH_JOBKEY_POLICIES) == 0) {
+ launch_data_dict_iterate(value, policy_setup, j);
+ }
+ break;
+ case 'k':
+ case 'K':
+ if (strcasecmp(key, LAUNCH_JOBKEY_KEEPALIVE) == 0) {
+ launch_data_dict_iterate(value, semaphoreitem_setup, j);
+ }
+ break;
+ case 'i':
+ case 'I':
+ if (strcasecmp(key, LAUNCH_JOBKEY_INETDCOMPATIBILITY) == 0) {
+ j->inetcompat = true;
+ j->abandon_pg = true;
+ if ((tmp = launch_data_dict_lookup(value, LAUNCH_JOBINETDCOMPATIBILITY_WAIT))) {
+ j->inetcompat_wait = launch_data_get_bool(tmp);
+ }
+ }
+ break;
+ case 'j':
+ case 'J':
+ if (strcasecmp(key, LAUNCH_JOBKEY_JETSAMPROPERTIES) == 0) {
+ launch_data_dict_iterate(value, (void (*)(launch_data_t, const char *, void *))jetsam_property_setup, j);
+ }
+ case 'e':
case 'E':
- if (strcasecmp(key, LAUNCH_JOBKEY_ENVIRONMENTVARIABLES) == 0)
+ if (strcasecmp(key, LAUNCH_JOBKEY_ENVIRONMENTVARIABLES) == 0) {
launch_data_dict_iterate(value, envitem_setup, j);
+ }
break;
case 'u':
case 'U':
calendarinterval_new_from_obj(j, value);
} else if (strcasecmp(key, LAUNCH_JOBKEY_SOFTRESOURCELIMITS) == 0) {
launch_data_dict_iterate(value, limititem_setup, j);
+#if HAVE_SANDBOX
+ } else if (strcasecmp(key, LAUNCH_JOBKEY_SANDBOXFLAGS) == 0) {
+ launch_data_dict_iterate(value, seatbelt_setup_flags, j);
+#endif
}
break;
case 'h':
case 'M':
if (strcasecmp(key, LAUNCH_JOBKEY_MACHSERVICES) == 0) {
launch_data_dict_iterate(value, machservice_setup, j);
- if (!SLIST_EMPTY(&j->machservices))
- job_setup_machport(j);
+ }
+ break;
+ case 'l':
+ case 'L':
+ if (strcasecmp(key, LAUNCH_JOBKEY_LAUNCHEVENTS) == 0) {
+ launch_data_dict_iterate(value, eventsystem_setup, j);
+ } else {
+ if (strcasecmp(key, LAUNCH_JOBKEY_LIMITLOADTOHARDWARE) == 0) {
+ return;
+ }
+ if (strcasecmp(key, LAUNCH_JOBKEY_LIMITLOADFROMHARDWARE) == 0) {
+ return;
+ }
}
break;
default:
+ job_log(j, LOG_WARNING, "Unknown key for dictionary: %s", key);
break;
}
}
void
-job_import_array(struct jobcb *j, const char *key, launch_data_t value)
+job_import_array(job_t j, const char *key, launch_data_t value)
{
- bool is_q_dir = false;
- bool is_wp = false;
+ size_t i, value_cnt = launch_data_array_get_count(value);
+ const char *str;
switch (key[0]) {
+ case 'p':
+ case 'P':
+ if (strcasecmp(key, LAUNCH_JOBKEY_PROGRAMARGUMENTS) == 0) {
+ return;
+ }
+ break;
+ case 'l':
+ case 'L':
+ if (strcasecmp(key, LAUNCH_JOBKEY_LIMITLOADTOHOSTS) == 0) {
+ return;
+ } else if (strcasecmp(key, LAUNCH_JOBKEY_LIMITLOADFROMHOSTS) == 0) {
+ return;
+ } else if (strcasecmp(key, LAUNCH_JOBKEY_LIMITLOADTOSESSIONTYPE) == 0) {
+ job_log(j, LOG_NOTICE, "launchctl should have transformed the \"%s\" array to a string", LAUNCH_JOBKEY_LIMITLOADTOSESSIONTYPE);
+ return;
+ }
+ break;
case 'q':
case 'Q':
if (strcasecmp(key, LAUNCH_JOBKEY_QUEUEDIRECTORIES) == 0) {
- is_q_dir = true;
- is_wp = true;
+ for (i = 0; i < value_cnt; i++) {
+ str = launch_data_get_string(launch_data_array_get_index(value, i));
+ if (job_assumes(j, str != NULL)) {
+ semaphoreitem_new(j, DIR_NOT_EMPTY, str);
+ }
+ }
+
}
break;
case 'w':
case 'W':
- if (strcasecmp(key, LAUNCH_JOBKEY_WATCHPATHS) == 0)
- is_wp = true;
+ if (strcasecmp(key, LAUNCH_JOBKEY_WATCHPATHS) == 0) {
+ for (i = 0; i < value_cnt; i++) {
+ str = launch_data_get_string(launch_data_array_get_index(value, i));
+ if (job_assumes(j, str != NULL)) {
+ semaphoreitem_new(j, PATH_CHANGES, str);
+ }
+ }
+ }
break;
case 'b':
case 'B':
- if (strcasecmp(key, LAUNCH_JOBKEY_BONJOURFDS) == 0)
+ if (strcasecmp(key, LAUNCH_JOBKEY_BONJOURFDS) == 0) {
socketgroup_setup(value, LAUNCH_JOBKEY_BONJOURFDS, j);
+ } else if (strcasecmp(key, LAUNCH_JOBKEY_BINARYORDERPREFERENCE) == 0) {
+ if (job_assumes(j, j->j_binpref = malloc(value_cnt * sizeof(*j->j_binpref)))) {
+ j->j_binpref_cnt = value_cnt;
+ for (i = 0; i < value_cnt; i++) {
+ j->j_binpref[i] = (cpu_type_t) launch_data_get_integer(launch_data_array_get_index(value, i));
+ }
+ }
+ }
break;
case 's':
case 'S':
if (strcasecmp(key, LAUNCH_JOBKEY_STARTCALENDARINTERVAL) == 0) {
- size_t i = 0, ci_cnt = launch_data_array_get_count(value);
- for (i = 0; i < ci_cnt; i++)
+ for (i = 0; i < value_cnt; i++) {
calendarinterval_new_from_obj(j, launch_data_array_get_index(value, i));
+ }
}
break;
default:
+ job_log(j, LOG_WARNING, "Unknown key for array: %s", key);
break;
}
-
- if (is_wp) {
- size_t i, wp_cnt = launch_data_array_get_count(value);
- const char *thepath;
- for (i = 0; i < wp_cnt; i++) {
- thepath = launch_data_get_string(launch_data_array_get_index(value, i));
- watchpath_new(j, thepath, is_q_dir);
- }
- }
}
void
job_import_keys(launch_data_t obj, const char *key, void *context)
{
- struct jobcb *j = context;
+ job_t j = context;
launch_data_type_t kind;
- if (obj == NULL)
+ if (!launchd_assumes(obj != NULL)) {
return;
+ }
kind = launch_data_get_type(obj);
case LAUNCH_DATA_ARRAY:
job_import_array(j, key, obj);
break;
+ case LAUNCH_DATA_OPAQUE:
+ job_import_opaque(j, key, obj);
+ break;
default:
job_log(j, LOG_WARNING, "Unknown value type '%d' for key: %s", kind, key);
break;
}
}
-struct jobcb *
-job_import2(launch_data_t pload)
+job_t
+jobmgr_import2(jobmgr_t jm, launch_data_t pload)
{
launch_data_t tmp, ldpa;
const char *label = NULL, *prog = NULL;
const char **argv = NULL;
- struct jobcb *j;
+ job_t j;
- if (pload == NULL)
+ if (!jobmgr_assumes(jm, pload != NULL)) {
+ errno = EINVAL;
return NULL;
+ }
- if (launch_data_get_type(pload) != LAUNCH_DATA_DICTIONARY)
+ if (unlikely(launch_data_get_type(pload) != LAUNCH_DATA_DICTIONARY)) {
+ errno = EINVAL;
return NULL;
-
- if ((tmp = launch_data_dict_lookup(pload, LAUNCH_JOBKEY_LABEL)) &&
- (launch_data_get_type(tmp) == LAUNCH_DATA_STRING)) {
- label = launch_data_get_string(tmp);
}
- if ((tmp = launch_data_dict_lookup(pload, LAUNCH_JOBKEY_PROGRAM)) &&
- (launch_data_get_type(tmp) == LAUNCH_DATA_STRING)) {
- prog = launch_data_get_string(tmp);
+
+ if (unlikely(!(tmp = launch_data_dict_lookup(pload, LAUNCH_JOBKEY_LABEL)))) {
+ errno = EINVAL;
+ return NULL;
}
- ldpa = launch_data_dict_lookup(pload, LAUNCH_JOBKEY_PROGRAMARGUMENTS);
- if (label == NULL) {
+ if (unlikely(launch_data_get_type(tmp) != LAUNCH_DATA_STRING)) {
errno = EINVAL;
return NULL;
- } else if ((j = job_find(root_job, label)) != NULL) {
- errno = EEXIST;
+ }
+
+ if (unlikely(!(label = launch_data_get_string(tmp)))) {
+ errno = EINVAL;
return NULL;
- } else if (label[0] == '\0' || (strncasecmp(label, "", strlen("com.apple.launchd")) == 0) ||
- (strtol(label, NULL, 10) != 0)) {
- syslog(LOG_ERR, "Somebody attempted to use a reserved prefix for a label: %s", label);
- /* the empty string, com.apple.launchd and number prefixes for labels are reserved */
+ }
+
+#if TARGET_OS_EMBEDDED
+ if (unlikely(g_embedded_privileged_action && s_embedded_privileged_job)) {
+ if (unlikely(!(tmp = launch_data_dict_lookup(pload, LAUNCH_JOBKEY_USERNAME)))) {
+ errno = EPERM;
+ return NULL;
+ }
+
+ const char *username = NULL;
+ if (likely(tmp && launch_data_get_type(tmp) == LAUNCH_DATA_STRING)) {
+ username = launch_data_get_string(tmp);
+ } else {
+ errno = EPERM;
+ return NULL;
+ }
+
+ if (!jobmgr_assumes(jm, s_embedded_privileged_job->username != NULL && username != NULL)) {
+ errno = EPERM;
+ return NULL;
+ }
+
+ if (unlikely(strcmp(s_embedded_privileged_job->username, username) != 0)) {
+ errno = EPERM;
+ return NULL;
+ }
+ } else if (g_embedded_privileged_action) {
errno = EINVAL;
return NULL;
}
+#endif
+
+ if ((tmp = launch_data_dict_lookup(pload, LAUNCH_JOBKEY_PROGRAM)) &&
+ (launch_data_get_type(tmp) == LAUNCH_DATA_STRING)) {
+ prog = launch_data_get_string(tmp);
+ }
+
+ int argc = 0;
+ if ((ldpa = launch_data_dict_lookup(pload, LAUNCH_JOBKEY_PROGRAMARGUMENTS))) {
+ size_t i, c;
+
+ if (launch_data_get_type(ldpa) != LAUNCH_DATA_ARRAY) {
+ errno = EINVAL;
+ return NULL;
+ }
- if (ldpa) {
- size_t i, c = launch_data_array_get_count(ldpa);
+ c = launch_data_array_get_count(ldpa);
argv = alloca((c + 1) * sizeof(char *));
- for (i = 0; i < c; i++)
- argv[i] = launch_data_get_string(launch_data_array_get_index(ldpa, i));
+ for (i = 0; i < c; i++) {
+ tmp = launch_data_array_get_index(ldpa, i);
+
+ if (launch_data_get_type(tmp) != LAUNCH_DATA_STRING) {
+ errno = EINVAL;
+ return NULL;
+ }
+
+ argv[i] = launch_data_get_string(tmp);
+ }
+
argv[i] = NULL;
+ argc = i;
+ }
+
+ if (!prog && argc == 0) {
+ jobmgr_log(jm, LOG_ERR, "Job specifies neither Program nor ProgramArguments: %s", label);
+ errno = EINVAL;
+ return NULL;
+ }
+
+ /* Find the requested session. You cannot load services into XPC domains in
+ * this manner.
+ */
+ launch_data_t session = launch_data_dict_lookup(pload, LAUNCH_JOBKEY_LIMITLOADTOSESSIONTYPE);
+ if (session) {
+ jobmgr_t jmt = NULL;
+ if (launch_data_get_type(session) == LAUNCH_DATA_STRING) {
+ jmt = jobmgr_find_by_name(jm, launch_data_get_string(session));
+ if (!jmt) {
+ jobmgr_log(jm, LOG_ERR, "Could not find requested session: %s", launch_data_get_string(session));
+ } else {
+ jm = jmt;
+ }
+ } else {
+ jobmgr_log(jm, LOG_ERR, "Session type is not a string.");
+ }
+
+ if (!jmt) {
+ errno = EINVAL;
+ return NULL;
+ }
+ }
+
+ /* For legacy reasons, we have a global hash of all labels in all job
+ * managers. So rather than make it a global, we store it in the root job
+ * manager. But for an XPC domain, we store a local hash of all services in
+ * the domain.
+ */
+ jobmgr_t where2look = (jm->properties & BOOTSTRAP_PROPERTY_XPC_DOMAIN) ? jm : root_jobmgr;
+ if (unlikely((j = job_find(where2look, label)) != NULL)) {
+ if (jm->xpc_singleton) {
+ /* There can (and probably will be) multiple attemtps to import the
+ * same XPC service from the same framework. This is okay. It's
+ * treated as a singleton, so just return the existing one so that
+ * it may be aliased into the requesting process' XPC domain.
+ */
+ return j;
+ } else {
+ /* If we're not a global XPC domain, then it's an error to try
+ * importing the same job/service multiple times.
+ */
+ errno = EEXIST;
+ return NULL;
+ }
+ } else if (unlikely(!jobmgr_label_test(where2look, label))) {
+ errno = EINVAL;
+ return NULL;
}
+ jobmgr_log(jm, LOG_DEBUG, "Importing %s.", label);
- if ((j = job_new(root_job, label, prog, argv, NULL, MACH_PORT_NULL)))
+ if (likely(j = job_new(jm, label, prog, argv))) {
launch_data_dict_iterate(pload, job_import_keys, j);
+ if (!uuid_is_null(j->expected_audit_uuid)) {
+ uuid_string_t uuid_str;
+ uuid_unparse(j->expected_audit_uuid, uuid_str);
+ job_log(j, LOG_DEBUG, "Imported job. Waiting for session for UUID %s.", uuid_str);
+ LIST_INSERT_HEAD(&s_needing_sessions, j, needing_session_sle);
+ errno = ENEEDAUTH;
+ } else {
+ job_log(j, LOG_DEBUG, "No security session specified.");
+ j->asport = MACH_PORT_NULL;
+ }
+
+ if (j->event_monitor) {
+ if (job_assumes(j, LIST_FIRST(&j->events) == NULL)) {
+ struct machservice *msi = NULL;
+ SLIST_FOREACH(msi, &j->machservices, sle) {
+ if (msi->event_update_port) {
+ break;
+ }
+ }
+
+ if (job_assumes(j, msi != NULL)) {
+ /* Create our send-once right so we can kick things off. */
+ (void)job_assumes(j, launchd_mport_make_send_once(msi->port, &_s_event_update_port) == KERN_SUCCESS);
+ if (!LIST_EMPTY(&_s_event_systems)) {
+ eventsystem_ping();
+ }
+ }
+ } else {
+ job_log(j, LOG_ERR, "The event monitor job may not have a LaunchEvents dictionary.");
+ job_remove(j);
+ j = NULL;
+ }
+ }
+ }
return j;
}
-struct jobcb *
-job_find(struct jobcb *j, const char *label)
+bool
+jobmgr_label_test(jobmgr_t jm, const char *str)
{
- struct jobcb *jr, *ji;
+ char *endstr = NULL;
+ const char *ptr;
+
+ if (str[0] == '\0') {
+ jobmgr_log(jm, LOG_ERR, "Empty job labels are not allowed");
+ return false;
+ }
+
+ for (ptr = str; *ptr; ptr++) {
+ if (iscntrl(*ptr)) {
+ jobmgr_log(jm, LOG_ERR, "ASCII control characters are not allowed in job labels. Index: %td Value: 0x%hhx", ptr - str, *ptr);
+ return false;
+ }
+ }
- if (label[0] == '\0')
- return root_job;
+ strtoll(str, &endstr, 0);
- if (strcmp(j->label, label) == 0)
- return j;
+ if (str != endstr) {
+ jobmgr_log(jm, LOG_ERR, "Job labels are not allowed to begin with numbers: %s", str);
+ return false;
+ }
- SLIST_FOREACH(ji, &j->jobs, sle) {
- if ((jr = job_find(ji, label)))
- return jr;
+ if ((strncasecmp(str, "com.apple.launchd", strlen("com.apple.launchd")) == 0) ||
+ (strncasecmp(str, "com.apple.launchctl", strlen("com.apple.launchctl")) == 0)) {
+ jobmgr_log(jm, LOG_ERR, "Job labels are not allowed to use a reserved prefix: %s", str);
+ return false;
}
- errno = ESRCH;
- return NULL;
+ return true;
}
-struct jobcb *
-job_find_by_pid(struct jobcb *j, pid_t p)
+job_t
+job_find(jobmgr_t jm, const char *label)
{
- struct jobcb *jr, *ji;
-
- if (j->p == p)
- return j;
+ job_t ji;
+
+ if (!jm) {
+ jm = root_jobmgr;
+ }
+
+ LIST_FOREACH(ji, &jm->label_hash[hash_label(label)], label_hash_sle) {
+ if (unlikely(ji->removal_pending || ji->mgr->shutting_down)) {
+ continue; /* 5351245 and 5488633 respectively */
+ }
- SLIST_FOREACH(ji, &j->jobs, sle) {
- if ((jr = job_find_by_pid(ji, p)))
- return jr;
+ if (strcmp(ji->label, label) == 0) {
+ return ji;
+ }
}
errno = ESRCH;
return NULL;
}
-void
-job_export_all2(struct jobcb *j, launch_data_t where)
+/* Should try and consolidate with job_mig_intran2() and jobmgr_find_by_pid(). */
+job_t
+jobmgr_find_by_pid_deep(jobmgr_t jm, pid_t p, bool anon_okay)
{
- launch_data_t tmp;
- struct jobcb *ji;
+ job_t ji = NULL;
+ LIST_FOREACH(ji, &jm->active_jobs[ACTIVE_JOB_HASH(p)], pid_hash_sle) {
+ if (ji->p == p && (!ji->anonymous || (ji->anonymous && anon_okay)) ) {
+ return ji;
+ }
+ }
- if (launchd_assumes((tmp = job_export2(j, false)) != NULL))
- launch_data_dict_insert(where, tmp, j->label);
+ jobmgr_t jmi = NULL;
+ SLIST_FOREACH(jmi, &jm->submgrs, sle) {
+ if ((ji = jobmgr_find_by_pid_deep(jmi, p, anon_okay))) {
+ break;
+ }
+ }
- SLIST_FOREACH(ji, &j->jobs, sle)
- job_export_all2(ji, where);
+ return ji;
}
-launch_data_t
-job_export_all(void)
+job_t
+jobmgr_find_by_pid(jobmgr_t jm, pid_t p, bool create_anon)
{
- launch_data_t resp = launch_data_alloc(LAUNCH_DATA_DICTIONARY);
+ job_t ji;
- job_export_all2(root_job, resp);
+ LIST_FOREACH(ji, &jm->active_jobs[ACTIVE_JOB_HASH(p)], pid_hash_sle) {
+ if (ji->p == p) {
+ return ji;
+ }
+ }
- return resp;
+ return create_anon ? job_new_anonymous(jm, p) : NULL;
}
-void
-job_reap(struct jobcb *j)
+job_t
+job_mig_intran2(jobmgr_t jm, mach_port_t mport, pid_t upid)
{
- struct rusage ru;
- time_t td = time(NULL) - j->start_time;
- bool bad_exit = false;
- int status;
-
- job_log(j, LOG_DEBUG, "Reaping");
+ jobmgr_t jmi;
+ job_t ji;
- if (j->execfd) {
- launchd_assumes(close(j->execfd) == 0);
- j->execfd = 0;
+ if (jm->jm_port == mport) {
+ return jobmgr_find_by_pid(jm, upid, true);
}
- if (!launchd_assumes(wait4(j->p, &status, 0, &ru) != -1)) {
- return;
+ SLIST_FOREACH(jmi, &jm->submgrs, sle) {
+ job_t jr;
+
+ if ((jr = job_mig_intran2(jmi, mport, upid))) {
+ return jr;
+ }
}
- if (j->wait_reply_port) {
- job_log(j, LOG_DEBUG, "MPM wait reply being sent");
- launchd_assumes(mpm_wait_reply(j->wait_reply_port, 0, status) == 0);
- j->wait_reply_port = MACH_PORT_NULL;
+ LIST_FOREACH(ji, &jm->jobs, sle) {
+ if (ji->j_port == mport) {
+ return ji;
+ }
}
- timeradd(&ru.ru_utime, &j->ru.ru_utime, &j->ru.ru_utime);
- timeradd(&ru.ru_stime, &j->ru.ru_stime, &j->ru.ru_stime);
- j->ru.ru_maxrss += ru.ru_maxrss;
- j->ru.ru_ixrss += ru.ru_ixrss;
- j->ru.ru_idrss += ru.ru_idrss;
- j->ru.ru_isrss += ru.ru_isrss;
- j->ru.ru_minflt += ru.ru_minflt;
- j->ru.ru_majflt += ru.ru_majflt;
- j->ru.ru_nswap += ru.ru_nswap;
- j->ru.ru_inblock += ru.ru_inblock;
- j->ru.ru_oublock += ru.ru_oublock;
- j->ru.ru_msgsnd += ru.ru_msgsnd;
- j->ru.ru_msgrcv += ru.ru_msgrcv;
- j->ru.ru_nsignals += ru.ru_nsignals;
- j->ru.ru_nvcsw += ru.ru_nvcsw;
- j->ru.ru_nivcsw += ru.ru_nivcsw;
+ return NULL;
+}
- if (WIFEXITED(status) && WEXITSTATUS(status) != 0) {
- job_log(j, LOG_WARNING, "exited with exit code: %d", WEXITSTATUS(status));
- bad_exit = true;
- }
+job_t
+job_mig_intran(mach_port_t p)
+{
+ struct ldcred *ldc = runtime_get_caller_creds();
+ job_t jr;
- if (WIFSIGNALED(status)) {
- int s = WTERMSIG(status);
- if (SIGKILL == s || SIGTERM == s) {
- job_log(j, LOG_NOTICE, "Exited: %s", strsignal(s));
- } else {
- job_log(j, LOG_WARNING, "Exited abnormally: %s", strsignal(s));
- bad_exit = true;
- }
- }
+ jr = job_mig_intran2(root_jobmgr, p, ldc->pid);
- if (!j->ondemand && !j->legacy_mach_job) {
- if (td < LAUNCHD_MIN_JOB_RUN_TIME) {
- job_log(j, LOG_WARNING, "respawning too quickly! throttling");
- bad_exit = true;
- j->throttle = true;
- } else if (td >= LAUNCHD_REWARD_JOB_RUN_TIME) {
- job_log(j, LOG_INFO, "lived long enough, forgiving past exit failures");
- j->failed_exits = 0;
+ if (!jobmgr_assumes(root_jobmgr, jr != NULL)) {
+ struct proc_bsdshortinfo proc;
+ if (proc_pidinfo(ldc->pid, PROC_PIDT_SHORTBSDINFO, 1, &proc, PROC_PIDT_SHORTBSDINFO_SIZE) == 0) {
+ if (errno != ESRCH) {
+ (void)jobmgr_assumes(root_jobmgr, errno == 0);
+ } else {
+ jobmgr_log(root_jobmgr, LOG_ERR, "%s() was confused by PID %u UID %u EUID %u Mach Port 0x%x: %s", __func__, ldc->pid, ldc->uid, ldc->euid, p, proc.pbsi_comm);
+ }
}
}
- if (!j->legacy_mach_job && bad_exit)
- j->failed_exits++;
+ return jr;
+}
+
+job_t
+job_find_by_service_port(mach_port_t p)
+{
+ struct machservice *ms;
- if (j->failed_exits > 0) {
- int failures_left = LAUNCHD_FAILED_EXITS_THRESHOLD - j->failed_exits;
- if (failures_left)
- job_log(j, LOG_WARNING, "%d more failure%s without living at least %d seconds will cause job removal",
- failures_left, failures_left > 1 ? "s" : "", LAUNCHD_REWARD_JOB_RUN_TIME);
+ LIST_FOREACH(ms, &port_hash[HASH_PORT(p)], port_hash_sle) {
+ if (ms->recv && (ms->port == p)) {
+ return ms->job;
+ }
}
- total_children--;
- j->last_exit_status = status;
- j->p = 0;
+ return NULL;
}
void
-job_dispatch(struct jobcb *j)
+job_mig_destructor(job_t j)
{
- if (job_active(j)) {
- return;
- } else if (job_useless(j)) {
+ /*
+ * 5477111
+ *
+ * 'j' can be invalid at this point. We should fix this up after Leopard ships.
+ */
+
+ if (unlikely(j && (j != workaround_5477111) && j->unload_at_mig_return)) {
+ job_log(j, LOG_NOTICE, "Unloading PID %u at MIG return.", j->p);
job_remove(j);
- } else if (job_keepalive(j)) {
- job_start(j);
- } else {
- job_watch(j);
}
+
+ workaround_5477111 = NULL;
+
+ calendarinterval_sanity_check();
}
void
-job_callback(void *obj, struct kevent *kev)
+job_export_all2(jobmgr_t jm, launch_data_t where)
{
- struct jobcb *j = obj;
- bool d = j->debug;
- int oldmask = 0;
+ jobmgr_t jmi;
+ job_t ji;
- if (d) {
- oldmask = setlogmask(LOG_UPTO(LOG_DEBUG));
+ SLIST_FOREACH(jmi, &jm->submgrs, sle) {
+ job_export_all2(jmi, where);
}
- switch (kev->filter) {
- case EVFILT_PROC:
- job_reap(j);
+ LIST_FOREACH(ji, &jm->jobs, sle) {
+ launch_data_t tmp;
- if (j->firstborn) {
- job_log(j, LOG_DEBUG, "first born died, begin shutdown");
- launchd_shutdown();
- } else {
- job_dispatch(j);
- }
- break;
- case EVFILT_TIMER:
- if ((uintptr_t)j == kev->ident) {
- job_start(j);
- } else {
- calendarinterval_callback(j, kev);
+ if (jobmgr_assumes(jm, (tmp = job_export(ji)) != NULL)) {
+ launch_data_dict_insert(where, tmp, ji->label);
}
- break;
- case EVFILT_VNODE:
- watchpath_callback(j, kev);
- break;
- case EVFILT_READ:
- if ((int)kev->ident != j->execfd) {
- socketgroup_callback(j, kev);
- break;
- }
- if (j->wait4debugger) {
- /* Allow somebody else to attach */
- launchd_assumes(kill(j->p, SIGSTOP) != -1);
- launchd_assumes(ptrace(PT_DETACH, j->p, NULL, 0) != -1);
- }
- if (kev->data > 0) {
- int e;
-
- read(j->execfd, &e, sizeof(e));
- errno = e;
- job_log_error(j, LOG_ERR, "execve()");
- job_remove(j);
- j = NULL;
- } else {
- launchd_assumes(close(j->execfd) == 0);
- j->execfd = 0;
- }
- break;
- case EVFILT_MACHPORT:
- job_start(j);
- break;
- default:
- launchd_assumes(false);
- break;
}
+}
- if (d) {
- /* the job might have been removed, must not call job_log() */
- setlogmask(oldmask);
+launch_data_t
+job_export_all(void)
+{
+ launch_data_t resp = launch_data_alloc(LAUNCH_DATA_DICTIONARY);
+
+ if (launchd_assumes(resp != NULL)) {
+ job_export_all2(root_jobmgr, resp);
}
+
+ return resp;
}
void
-job_start(struct jobcb *j)
+job_log_stray_pg(job_t j)
{
- int spair[2];
- int execspair[2];
- char nbuf[64];
- pid_t c;
- bool sipc = false;
-
- if (!launchd_assumes(j->req_port == MACH_PORT_NULL))
+ pid_t *pids = NULL;
+ size_t len = sizeof(pid_t) * get_kern_max_proc();
+ int i = 0, kp_cnt = 0;
+
+ if (!do_apple_internal_logging) {
return;
+ }
- if (!launchd_assumes(j->parent != NULL))
- return;
+ runtime_ktrace(RTKT_LAUNCHD_FINDING_STRAY_PG, j->p, 0, 0);
- if (job_active(j)) {
- job_log(j, LOG_DEBUG, "Already started");
- return;
- } else if (!j->legacy_mach_job && j->throttle) {
- j->throttle = false;
- job_log(j, LOG_WARNING, "Throttling: Will restart in %d seconds", LAUNCHD_MIN_JOB_RUN_TIME);
- launchd_assumes(kevent_mod((uintptr_t)j, EVFILT_TIMER, EV_ADD|EV_ONESHOT,
- NOTE_SECONDS, LAUNCHD_MIN_JOB_RUN_TIME, j) != -1);
+ if (!job_assumes(j, (pids = malloc(len)) != NULL)) {
return;
}
-
- job_log(j, LOG_DEBUG, "Starting");
-
- if (!j->legacy_mach_job)
- sipc = (!SLIST_EMPTY(&j->sockets) || !SLIST_EMPTY(&j->machservices));
-
- /* FIXME, using stdinpath is a hack for re-reading the conf file */
- if (j->stdinpath)
- sipc = true;
-
- j->checkedin = false;
-
- if (sipc)
- socketpair(AF_UNIX, SOCK_STREAM, 0, spair);
-
- socketpair(AF_UNIX, SOCK_STREAM, 0, execspair);
-
- time(&j->start_time);
-
- if (j->bs_port) {
- launchd_assumes(launchd_mport_notify_req(j->bs_port, MACH_NOTIFY_NO_SENDERS) == KERN_SUCCESS);
+ if (!job_assumes(j, (kp_cnt = proc_listpgrppids(j->p, pids, len)) != -1)) {
+ goto out;
}
-
- switch (c = job_fork(j->bs_port ? j : j->parent)) {
- case -1:
- job_log_error(j, LOG_ERR, "fork() failed, will try again in one second");
- launchd_assumes(close(execspair[0]) == 0);
- launchd_assumes(close(execspair[1]) == 0);
- if (sipc) {
- launchd_assumes(close(spair[0]) == 0);
- launchd_assumes(close(spair[1]) == 0);
+
+ for (i = 0; i < kp_cnt; i++) {
+ pid_t p_i = pids[i];
+ if (p_i == j->p) {
+ continue;
+ } else if (!job_assumes(j, p_i != 0 && p_i != 1)) {
+ continue;
}
- break;
- case 0:
- launchd_assumes(close(execspair[0]) == 0);
- /* wait for our parent to say they've attached a kevent to us */
- read(_fd(execspair[1]), &c, sizeof(c));
- if (j->firstborn) {
- setpgid(getpid(), getpid());
- if (isatty(STDIN_FILENO)) {
- if (tcsetpgrp(STDIN_FILENO, getpid()) == -1)
- job_log_error(j, LOG_WARNING, "tcsetpgrp()");
+
+ struct proc_bsdshortinfo proc;
+ if (proc_pidinfo(p_i, PROC_PIDT_SHORTBSDINFO, 1, &proc, PROC_PIDT_SHORTBSDINFO_SIZE) == 0) {
+ if (errno != ESRCH) {
+ job_assumes(j, errno == 0);
}
+ continue;
}
+
+ pid_t pp_i = proc.pbsi_ppid;
+ const char *z = (proc.pbsi_status == SZOMB) ? "zombie " : "";
+ const char *n = proc.pbsi_comm;
- if (sipc) {
- launchd_assumes(close(spair[0]) == 0);
- sprintf(nbuf, "%d", spair[1]);
- setenv(LAUNCHD_TRUSTED_FD_ENV, nbuf, 1);
- }
- job_start_child(j, execspair[1]);
- break;
- default:
- if (!SLIST_EMPTY(&j->machservices))
- j->priv_port_has_senders = true;
- j->p = c;
- total_children++;
- launchd_assumes(close(execspair[1]) == 0);
- j->execfd = _fd(execspair[0]);
- if (sipc) {
- launchd_assumes(close(spair[1]) == 0);
- ipc_open(_fd(spair[0]), j);
- }
- if (kevent_mod(j->execfd, EVFILT_READ, EV_ADD, 0, 0, &j->kqjob_callback) == -1)
- job_log_error(j, LOG_ERR, "kevent_mod(j->execfd): %m");
- if (kevent_mod(c, EVFILT_PROC, EV_ADD, NOTE_EXIT, 0, &j->kqjob_callback) == -1) {
- job_log_error(j, LOG_ERR, "kevent()");
- job_reap(j);
- } else {
- if (j->ondemand)
- job_ignore(j);
- }
-
- if (!j->stall_before_exec) {
- /* this unblocks the child and avoids a race
- * between the above fork() and the kevent_mod() */
- write(j->execfd, &c, sizeof(c));
- }
- break;
+ job_log(j, LOG_WARNING, "Stray %sprocess with PGID equal to this dead job: PID %u PPID %u PGID %u %s", z, p_i, pp_i, proc.pbsi_pgid, n);
}
+
+out:
+ free(pids);
}
void
-job_start_child(struct jobcb *j, int execfd)
+job_reap(job_t j)
{
- const char *file2exec = "/usr/libexec/launchproxy";
- const char **argv;
- int gflags = GLOB_NOSORT|GLOB_NOCHECK|GLOB_TILDE|GLOB_DOOFFS;
- glob_t g;
- int i;
+ struct rusage ru;
+ int status;
- job_setup_attributes(j);
+ bool is_system_bootstrapper = j->is_bootstrapper && pid1_magic && !j->mgr->parentmgr;
+
+ job_log(j, LOG_DEBUG, "Reaping");
- if (j->argv && j->globargv) {
- g.gl_offs = 1;
- for (i = 0; i < j->argc; i++) {
- if (i > 0)
- gflags |= GLOB_APPEND;
- if (glob(j->argv[i], gflags, NULL, &g) != 0) {
- job_log_error(j, LOG_ERR, "glob(\"%s\")", j->argv[i]);
- exit(EXIT_FAILURE);
- }
- }
- g.gl_pathv[0] = (char *)file2exec;
- argv = (const char **)g.gl_pathv;
- } else if (j->argv) {
- argv = alloca((j->argc + 2) * sizeof(char *));
- argv[0] = file2exec;
- for (i = 0; i < j->argc; i++)
- argv[i + 1] = j->argv[i];
- argv[i + 1] = NULL;
- } else {
- argv = alloca(3 * sizeof(char *));
- argv[0] = file2exec;
- argv[1] = j->prog;
- argv[2] = NULL;
+ if (j->shmem) {
+ (void)job_assumes(j, vm_deallocate(mach_task_self(), (vm_address_t)j->shmem, getpagesize()) == 0);
+ j->shmem = NULL;
}
- if (!j->inetcompat)
- argv++;
-
- if (j->wait4debugger && ptrace(PT_TRACE_ME, getpid(), NULL, 0) == -1)
- job_log_error(j, LOG_ERR, "ptrace(PT_TRACE_ME, ...)");
+ if (unlikely(j->weird_bootstrap)) {
+ int64_t junk = 0;
+ job_mig_swap_integer(j, VPROC_GSK_WEIRD_BOOTSTRAP, 0, 0, &junk);
+ }
- if (j->force_ppc) {
- int affinmib[] = { CTL_KERN, KERN_AFFINITY, 1, 1 };
- size_t mibsz = sizeof(affinmib) / sizeof(affinmib[0]);
+ if (j->log_redirect_fd && !j->legacy_LS_job) {
+ job_log_stdouterr(j); /* one last chance */
- if (sysctl(affinmib, mibsz, NULL, NULL, NULL, 0) == -1)
- job_log_error(j, LOG_WARNING, "Failed to force PowerPC execution");
+ if (j->log_redirect_fd) {
+ (void)job_assumes(j, runtime_close(j->log_redirect_fd) != -1);
+ j->log_redirect_fd = 0;
+ }
}
- if (j->prog) {
- execv(j->inetcompat ? file2exec : j->prog, (char *const*)argv);
- job_log_error(j, LOG_ERR, "execv(\"%s\", ...)", j->prog);
- } else {
- execvp(j->inetcompat ? file2exec : argv[0], (char *const*)argv);
- job_log_error(j, LOG_ERR, "execvp(\"%s\", ...)", argv[0]);
+ if (j->fork_fd) {
+ (void)job_assumes(j, runtime_close(j->fork_fd) != -1);
+ j->fork_fd = 0;
}
- write(execfd, &errno, sizeof(errno));
- exit(EXIT_FAILURE);
-}
+ if (j->anonymous) {
+ status = 0;
+ memset(&ru, 0, sizeof(ru));
+ } else {
+ /*
+ * The job is dead. While the PID/PGID is still known to be
+ * valid, try to kill abandoned descendant processes.
+ */
+ job_log_stray_pg(j);
+ if (!j->abandon_pg) {
+ if (unlikely(runtime_killpg(j->p, SIGTERM) == -1 && errno != ESRCH)) {
+#ifdef __LP64__
+ job_log(j, LOG_APPLEONLY, "Bug: 5487498");
+#else
+ (void)job_assumes(j, false);
+#endif
+ }
+ }
-void job_setup_env_from_other_jobs(struct jobcb *j)
-{
- struct envitem *ei;
- struct jobcb *ji;
+ /* We have to work around one of two kernel bugs here. ptrace(3) may
+ * have abducted the child away from us and reparented it to the tracing
+ * process. If the process then exits, we still get NOTE_EXIT, but we
+ * cannot reap it because the kernel may not have restored the true
+ * parent/child relationship in time.
+ *
+ * See <rdar://problem/5020256>.
+ *
+ * The other bug is if the shutdown monitor has suspended a task and not
+ * resumed it before exiting. In this case, the kernel will not clean up
+ * after the shutdown monitor. It will, instead, leave the task
+ * task suspended and not process any pending signals on the event loop
+ * for the task.
+ *
+ * There are a variety of other kernel bugs that could prevent a process
+ * from exiting, usually having to do with faulty hardware or talking to
+ * misbehaving drivers that mark a thread as uninterruptible and
+ * deadlock/hang before unmarking it as such. So we have to work around
+ * that too.
+ *
+ * See <rdar://problem/9284889&9359725>.
+ */
+ if (j->workaround9359725) {
+ job_log(j, LOG_NOTICE, "Simulated exit: <rdar://problem/9359725>");
+ status = W_EXITCODE(-1, SIGSEGV);
+ memset(&ru, 0, sizeof(ru));
+ } else if (wait4(j->p, &status, 0, &ru) == -1) {
+ job_log(j, LOG_NOTICE, "Assuming job exited: <rdar://problem/5020256>: %d: %s", errno, strerror(errno));
+ status = W_EXITCODE(-1, SIGSEGV);
+ memset(&ru, 0, sizeof(ru));
+ }
+ }
- SLIST_FOREACH(ji, &j->jobs, sle)
- job_setup_env_from_other_jobs(ji);
+ if (j->exit_timeout) {
+ kevent_mod((uintptr_t)&j->exit_timeout, EVFILT_TIMER, EV_DELETE, 0, 0, NULL);
+ }
+
+ LIST_REMOVE(j, pid_hash_sle);
- SLIST_FOREACH(ei, &j->global_env, sle)
- setenv(ei->key, ei->value, 1);
-}
+ if (j->sent_signal_time) {
+ uint64_t td_sec, td_usec, td = runtime_get_nanoseconds_since(j->sent_signal_time);
-void
-job_setup_attributes(struct jobcb *j)
-{
- struct limititem *li;
- struct envitem *ei;
- struct group *gre = NULL;
- gid_t gre_g = 0;
+ td_sec = td / NSEC_PER_SEC;
+ td_usec = (td % NSEC_PER_SEC) / NSEC_PER_USEC;
- setpriority(PRIO_PROCESS, 0, j->nice);
+ job_log(j, LOG_DEBUG, "Exited %llu.%06llu seconds after the first signal was sent", td_sec, td_usec);
+ }
- SLIST_FOREACH(li, &j->limits, sle) {
- struct rlimit rl;
+ timeradd(&ru.ru_utime, &j->ru.ru_utime, &j->ru.ru_utime);
+ timeradd(&ru.ru_stime, &j->ru.ru_stime, &j->ru.ru_stime);
+ j->ru.ru_maxrss += ru.ru_maxrss;
+ j->ru.ru_ixrss += ru.ru_ixrss;
+ j->ru.ru_idrss += ru.ru_idrss;
+ j->ru.ru_isrss += ru.ru_isrss;
+ j->ru.ru_minflt += ru.ru_minflt;
+ j->ru.ru_majflt += ru.ru_majflt;
+ j->ru.ru_nswap += ru.ru_nswap;
+ j->ru.ru_inblock += ru.ru_inblock;
+ j->ru.ru_oublock += ru.ru_oublock;
+ j->ru.ru_msgsnd += ru.ru_msgsnd;
+ j->ru.ru_msgrcv += ru.ru_msgrcv;
+ j->ru.ru_nsignals += ru.ru_nsignals;
+ j->ru.ru_nvcsw += ru.ru_nvcsw;
+ j->ru.ru_nivcsw += ru.ru_nivcsw;
- if (getrlimit(li->which, &rl) == -1) {
- job_log_error(j, LOG_WARNING, "getrlimit()");
- continue;
+ if (WIFEXITED(status) && WEXITSTATUS(status) != 0) {
+ int level = LOG_WARNING;
+ if (!j->did_exec && (j->fail_cnt++ % LAUNCHD_LOG_FAILED_EXEC_FREQ) != 0) {
+ level = LOG_DEBUG;
}
- if (li->sethard)
- rl.rlim_max = li->lim.rlim_max;
- if (li->setsoft)
- rl.rlim_cur = li->lim.rlim_cur;
-
- if (setrlimit(li->which, &rl) == -1)
- job_log_error(j, LOG_WARNING, "setrlimit()");
+ job_log(j, level, "Exited with code: %d", WEXITSTATUS(status));
+ } else {
+ j->fail_cnt = 0;
}
- if (!j->inetcompat && j->session_create)
- launchd_SessionCreate();
-
- if (j->low_pri_io) {
- int lowprimib[] = { CTL_KERN, KERN_PROC_LOW_PRI_IO };
- int val = 1;
-
- if (sysctl(lowprimib, sizeof(lowprimib) / sizeof(lowprimib[0]), NULL, NULL, &val, sizeof(val)) == -1)
- job_log_error(j, LOG_WARNING, "sysctl(\"%s\")", "kern.proc_low_pri_io");
- }
- if (j->rootdir) {
- chroot(j->rootdir);
- chdir(".");
- }
- if (j->groupname) {
- gre = getgrnam(j->groupname);
- if (gre) {
- gre_g = gre->gr_gid;
- if (-1 == setgid(gre_g)) {
- job_log_error(j, LOG_ERR, "setgid(%d)", gre_g);
- exit(EXIT_FAILURE);
+ if (WIFSIGNALED(status)) {
+ int s = WTERMSIG(status);
+ if ((SIGKILL == s || SIGTERM == s) && !j->stopped) {
+ job_log(j, LOG_NOTICE, "Exited: %s", strsignal(s));
+ } else if (!j->stopped && !j->clean_kill) {
+ switch (s) {
+ /* Signals which indicate a crash. */
+ case SIGILL:
+ case SIGABRT:
+ case SIGFPE:
+ case SIGBUS:
+ case SIGSEGV:
+ case SIGSYS:
+ /* If the kernel has posted NOTE_EXIT and the signal sent to the process was
+ * SIGTRAP, assume that it's a crash.
+ */
+ case SIGTRAP:
+ j->crashed = true;
+ job_log(j, LOG_WARNING, "Job appears to have crashed: %s", strsignal(s));
+ break;
+ default:
+ job_log(j, LOG_WARNING, "Exited abnormally: %s", strsignal(s));
+ break;
+ }
+
+ if (is_system_bootstrapper && j->crashed) {
+ job_log(j, LOG_ERR | LOG_CONSOLE, "The %s bootstrapper has crashed: %s", j->mgr->name, strsignal(s));
}
- } else {
- job_log(j, LOG_ERR, "getgrnam(\"%s\") failed", j->groupname);
- exit(EXIT_FAILURE);
}
}
- if (j->username || j->mach_uid) {
- struct passwd *pwe;
-
- if (j->username)
- pwe = getpwnam(j->username);
- else
- pwe = getpwuid(j->mach_uid);
- if (pwe) {
- uid_t pwe_u = pwe->pw_uid;
- uid_t pwe_g = pwe->pw_gid;
-
- if (pwe->pw_expire && time(NULL) >= pwe->pw_expire) {
- job_log(j, LOG_ERR, "expired account: %s", j->username);
- exit(EXIT_FAILURE);
- }
- if (j->init_groups) {
- if (-1 == initgroups(j->username, gre ? gre_g : pwe_g)) {
- job_log_error(j, LOG_ERR, "initgroups()");
- exit(EXIT_FAILURE);
- }
- }
- if (!gre) {
- if (-1 == setgid(pwe_g)) {
- job_log_error(j, LOG_ERR, "setgid(%d)", pwe_g);
- exit(EXIT_FAILURE);
- }
- }
- if (-1 == setuid(pwe_u)) {
- job_log_error(j, LOG_ERR, "setuid(%d)", pwe_u);
- exit(EXIT_FAILURE);
+ j->reaped = true;
+
+ struct machservice *msi = NULL;
+ if (j->crashed || !(j->did_exec || j->anonymous)) {
+ SLIST_FOREACH(msi, &j->machservices, sle) {
+ if (j->crashed && !msi->isActive && (msi->drain_one_on_crash || msi->drain_all_on_crash)) {
+ machservice_drain_port(msi);
}
- } else {
- if (j->username) {
- job_log(j, LOG_WARNING, "getpwnam(\"%s\") failed", j->username);
- } else {
- job_log(j, LOG_WARNING, "getpwuid(\"%d\") failed", j->mach_uid);
+
+ if (!j->did_exec && msi->reset && job_assumes(j, !msi->isActive)) {
+ machservice_resetport(j, msi);
}
- exit(EXIT_FAILURE);
}
}
- if (j->workingdir)
- chdir(j->workingdir);
- if (j->setmask)
- umask(j->mask);
- if (j->stdinpath) {
- int sifd = open(j->stdinpath, O_RDONLY|O_NOCTTY);
- if (sifd == -1) {
- job_log_error(j, LOG_WARNING, "open(\"%s\", ...)", j->stdinpath);
- } else {
- launchd_assumes(dup2(sifd, STDIN_FILENO) != -1);
- launchd_assumes(close(sifd) == 0);
+
+ /* HACK: Essentially duplicating the logic directly above. But this has
+ * gotten really hairy, and I don't want to try consolidating it right now.
+ */
+ if (j->xpc_service && !j->xpcproxy_did_exec) {
+ job_log(j, LOG_ERR, "XPC Service could not exec(3). Resetting port.");
+ SLIST_FOREACH(msi, &j->machservices, sle) {
+ /* Drain the messages but do not reset the port. If xpcproxy could
+ * not exec(3), then we don't want to continue trying, since there
+ * is very likely a serious configuration error with the service.
+ *
+ * <rdar://problem/8986802>
+ */
+ machservice_resetport(j, msi);
}
}
- if (j->stdoutpath) {
- int sofd = open(j->stdoutpath, O_WRONLY|O_APPEND|O_CREAT|O_NOCTTY, DEFFILEMODE);
- if (sofd == -1) {
- job_log_error(j, LOG_WARNING, "open(\"%s\", ...)", j->stdoutpath);
- } else {
- launchd_assumes(dup2(sofd, STDOUT_FILENO) != -1);
- launchd_assumes(close(sofd) == 0);
+
+ struct suspended_peruser *spi = NULL;
+ while ((spi = LIST_FIRST(&j->suspended_perusers))) {
+ job_log(j, LOG_ERR, "Job exited before resuming per-user launchd for UID %u. Will forcibly resume.", spi->j->mach_uid);
+ spi->j->peruser_suspend_count--;
+ if (spi->j->peruser_suspend_count == 0) {
+ job_dispatch(spi->j, false);
}
+ LIST_REMOVE(spi, sle);
+ free(spi);
}
- if (j->stderrpath) {
- int sefd = open(j->stderrpath, O_WRONLY|O_APPEND|O_CREAT|O_NOCTTY, DEFFILEMODE);
- if (sefd == -1) {
- job_log_error(j, LOG_WARNING, "open(\"%s\", ...)", j->stderrpath);
- } else {
- launchd_assumes(dup2(sefd, STDERR_FILENO) != -1);
- launchd_assumes(close(sefd) == 0);
+
+ j->last_exit_status = status;
+
+ if (j->exit_status_dest) {
+ errno = helper_downcall_wait(j->exit_status_dest, j->last_exit_status);
+ if (errno && errno != MACH_SEND_INVALID_DEST) {
+ (void)job_assumes(j, errno == 0);
}
+
+ j->exit_status_dest = MACH_PORT_NULL;
}
- job_setup_env_from_other_jobs(root_job);
+ if (j->spawn_reply_port) {
+ /* If the child never called exec(3), we must send a spawn() reply so
+ * that the requestor can get exit status from it. If we fail to send
+ * the reply for some reason, we have to deallocate the exit status port
+ * ourselves.
+ */
+ kern_return_t kr = job_mig_spawn2_reply(j->spawn_reply_port, BOOTSTRAP_SUCCESS, j->p, j->exit_status_port);
+ if (kr) {
+ if (kr != MACH_SEND_INVALID_DEST) {
+ errno = kr;
+ (void)job_assumes(j, errno == KERN_SUCCESS);
+ }
- SLIST_FOREACH(ei, &j->env, sle)
- setenv(ei->key, ei->value, 1);
+ (void)job_assumes(j, launchd_mport_close_recv(j->exit_status_port) == KERN_SUCCESS);
+ }
- setsid();
-}
+ j->exit_status_port = MACH_PORT_NULL;
+ j->spawn_reply_port = MACH_PORT_NULL;
+ }
-int
-dir_has_files(const char *path)
-{
- DIR *dd = opendir(path);
- struct dirent *de;
- bool r = 0;
+ if (j->anonymous) {
+ total_anon_children--;
+ if (j->holds_ref) {
+ runtime_del_ref();
+ }
+ } else {
+ runtime_del_ref();
+ total_children--;
+ }
+
+ if (j->has_console) {
+ g_wsp = 0;
+ }
- if (!dd)
- return -1;
+ if (j->shutdown_monitor) {
+ job_log(j, LOG_NOTICE | LOG_CONSOLE, "Shutdown monitor has exited.");
+ _s_shutdown_monitor = NULL;
+ j->shutdown_monitor = false;
+ }
- while ((de = readdir(dd))) {
- if (strcmp(de->d_name, ".") && strcmp(de->d_name, "..")) {
- r = 1;
- break;
+ if (j->event_monitor && !j->mgr->shutting_down) {
+ msi = NULL;
+ SLIST_FOREACH(msi, &j->machservices, sle) {
+ if (msi->event_update_port) {
+ break;
+ }
+ }
+ /* Only do this if we've gotten the port-destroyed notification already.
+ * If we haven't yet, the port destruction handler will do this.
+ */
+ if (job_assumes(j, msi != NULL) && !msi->isActive) {
+ if (_s_event_update_port == MACH_PORT_NULL) {
+ (void)job_assumes(j, launchd_mport_make_send_once(msi->port, &_s_event_update_port) == KERN_SUCCESS);
+ }
+ eventsystem_ping();
}
}
-
- launchd_assumes(closedir(dd) == 0);
- return r;
+
+ if (!j->anonymous) {
+ j->mgr->normal_active_cnt--;
+ }
+ j->sent_signal_time = 0;
+ j->sent_sigkill = false;
+ j->clean_kill = false;
+ j->sent_kill_via_shmem = false;
+ j->lastlookup = NULL;
+ j->lastlookup_gennum = 0;
+ j->p = 0;
}
void
-calendarinterval_setalarm(struct jobcb *j, struct calendarinterval *ci)
+jobmgr_dispatch_all(jobmgr_t jm, bool newmounthack)
{
- time_t later;
+ jobmgr_t jmi, jmn;
+ job_t ji, jn;
- later = cronemu(ci->when.tm_mon, ci->when.tm_mday, ci->when.tm_hour, ci->when.tm_min);
+ if (jm->shutting_down) {
+ return;
+ }
- if (ci->when.tm_wday != -1) {
- time_t otherlater = cronemu_wday(ci->when.tm_wday, ci->when.tm_hour, ci->when.tm_min);
+ SLIST_FOREACH_SAFE(jmi, &jm->submgrs, sle, jmn) {
+ jobmgr_dispatch_all(jmi, newmounthack);
+ }
- if (ci->when.tm_mday == -1) {
- later = otherlater;
- } else {
- later = later < otherlater ? later : otherlater;
+ LIST_FOREACH_SAFE(ji, &jm->jobs, sle, jn) {
+ if (newmounthack && ji->start_on_mount) {
+ ji->start_pending = true;
}
- }
- if (-1 == kevent_mod((uintptr_t)ci, EVFILT_TIMER, EV_ADD, NOTE_ABSOLUTE|NOTE_SECONDS, later, j)) {
- job_log_error(j, LOG_ERR, "adding kevent alarm");
- } else {
- job_log(j, LOG_INFO, "scheduled to run again at %s", ctime(&later));
+ job_dispatch(ji, false);
}
}
-size_t
-job_prep_log_preface(struct jobcb *j, char *buf)
-{
- size_t lsz = strlen(j->label);
- char newlabel[lsz * 2 + 1];
- size_t i, o, r = 0;
-
- for (i = 0, o = 0; i < lsz; i++, o++) {
- if (j->label[i] == '%') {
- newlabel[o] = '%';
- o++;
- newlabel[o] = '%';
- } else {
- newlabel[o] = j->label[i];
+void
+job_dispatch_curious_jobs(job_t j)
+{
+ job_t ji = NULL, jt = NULL;
+ SLIST_FOREACH_SAFE(ji, &s_curious_jobs, curious_jobs_sle, jt) {
+ struct semaphoreitem *si = NULL;
+ SLIST_FOREACH(si, &ji->semaphores, sle) {
+ if (!(si->why == OTHER_JOB_ENABLED || si->why == OTHER_JOB_DISABLED)) {
+ continue;
+ }
+
+ if (strcmp(si->what, j->label) == 0) {
+ job_log(ji, LOG_DEBUG, "Dispatching out of interest in \"%s\".", j->label);
+
+ if (!ji->removing) {
+ job_dispatch(ji, false);
+ } else {
+ job_log(ji, LOG_NOTICE, "The following job is circularly dependent upon this one: %s", j->label);
+ }
+
+ /* ji could be removed here, so don't do anything with it or its semaphores
+ * after this point.
+ */
+ break;
+ }
}
}
- newlabel[o] = '\0';
-
- if (j->parent)
- r = job_prep_log_preface(j->parent, buf);
-
- return r + sprintf(buf + r, "%s%s", j->parent ? "/" : "", newlabel);
}
-void
-job_log_error(struct jobcb *j, int pri, const char *msg, ...)
+job_t
+job_dispatch(job_t j, bool kickstart)
{
- char newmsg[10000];
- va_list ap;
- size_t o;
+ /* Don't dispatch a job if it has no audit session set. */
+ if (!uuid_is_null(j->expected_audit_uuid)) {
+ return NULL;
+ }
+ if (j->alias) {
+ j = j->alias;
+ }
- o = job_prep_log_preface(j, newmsg);
+#if TARGET_OS_EMBEDDED
+ if (g_embedded_privileged_action && s_embedded_privileged_job) {
+ if (!job_assumes(j, s_embedded_privileged_job->username != NULL && j->username != NULL)) {
+ errno = EPERM;
+ return NULL;
+ }
+
+ if (strcmp(j->username, s_embedded_privileged_job->username) != 0) {
+ errno = EPERM;
+ return NULL;
+ }
+ } else if (g_embedded_privileged_action) {
+ errno = EINVAL;
+ return NULL;
+ }
+#endif
- sprintf(newmsg + o, ": %s: %s", msg, strerror(errno));
+ /*
+ * The whole job removal logic needs to be consolidated. The fact that
+ * a job can be removed from just about anywhere makes it easy to have
+ * stale pointers left behind somewhere on the stack that might get
+ * used after the deallocation. In particular, during job iteration.
+ *
+ * This is a classic example. The act of dispatching a job may delete it.
+ */
+ if (!job_active(j)) {
+ if (job_useless(j)) {
+ job_remove(j);
+ return NULL;
+ }
+ if (unlikely(j->per_user && j->peruser_suspend_count > 0)) {
+ return NULL;
+ }
+
+ if (kickstart || job_keepalive(j)) {
+ job_log(j, LOG_DEBUG, "Starting job (kickstart = %s)", kickstart ? "true" : "false");
+ job_start(j);
+ } else {
+ job_log(j, LOG_DEBUG, "Watching job (kickstart = %s)", kickstart ? "true" : "false");
+ job_watch(j);
+
+ /*
+ * 5455720
+ *
+ * Path checking and monitoring is really racy right now.
+ * We should clean this up post Leopard.
+ */
+ if (job_keepalive(j)) {
+ job_start(j);
+ }
+ }
+ } else {
+ job_log(j, LOG_DEBUG, "Tried to dispatch an already active job (%s).", job_active(j));
+ }
- va_start(ap, msg);
- vsyslog(pri, newmsg, ap);
- va_end(ap);
+ return j;
}
void
-job_log(struct jobcb *j, int pri, const char *msg, ...)
+job_log_stdouterr2(job_t j, const char *msg, ...)
{
- char newmsg[10000];
+ struct runtime_syslog_attr attr = { j->label, j->label, j->mgr->name, LOG_NOTICE, getuid(), j->p, j->p };
va_list ap;
- size_t o;
-
- o = job_prep_log_preface(j, newmsg);
-
- sprintf(newmsg + o, ": %s", msg);
va_start(ap, msg);
- vsyslog(pri, newmsg, ap);
+ runtime_vsyslog(&attr, msg, ap);
va_end(ap);
}
-bool
-watchpath_new(struct jobcb *j, const char *name, bool qdir)
-{
- struct watchpath *wp = calloc(1, sizeof(struct watchpath) + strlen(name) + 1);
-
- if (!launchd_assumes(wp != NULL))
- return false;
-
- wp->is_qdir = qdir;
-
- wp->fd = -1; /* watchpath_watch() will open this */
-
- strcpy(wp->name, name);
-
- SLIST_INSERT_HEAD(&j->vnodes, wp, sle);
-
- return true;
-}
-
-void
-watchpath_delete(struct jobcb *j, struct watchpath *wp)
-{
- if (wp->fd != -1)
- launchd_assumes(close(wp->fd) != -1);
-
- SLIST_REMOVE(&j->vnodes, wp, watchpath, sle);
-
- free(wp);
-}
-
-void
-watchpath_ignore(struct jobcb *j, struct watchpath *wp)
-{
- if (wp->fd != -1) {
- job_log(j, LOG_DEBUG, "Ignoring Vnode: %d", wp->fd);
- launchd_assumes(kevent_mod(wp->fd, EVFILT_VNODE, EV_DELETE, 0, 0, NULL) != -1);
- }
-}
-
void
-watchpath_watch(struct jobcb *j, struct watchpath *wp)
+job_log_stdouterr(job_t j)
{
- int fflags = NOTE_WRITE|NOTE_EXTEND|NOTE_ATTRIB|NOTE_LINK;
- int qdir_file_cnt;
-
- if (!wp->is_qdir)
- fflags |= NOTE_DELETE|NOTE_RENAME|NOTE_REVOKE;
-
- if (wp->fd == -1)
- wp->fd = _fd(open(wp->name, O_EVTONLY|O_NOCTTY|O_NOFOLLOW));
+ char *msg, *bufindex, *buf = malloc(BIG_PIPE_SIZE + 1);
+ bool close_log_redir = false;
+ ssize_t rsz;
- if (wp->fd == -1)
- return job_log_error(j, LOG_ERR, "Watchpath monitoring failed on \"%s\"", wp->name);
-
- job_log(j, LOG_DEBUG, "Watching Vnode: %d", wp->fd);
- launchd_assumes(kevent_mod(wp->fd, EVFILT_VNODE, EV_ADD|EV_CLEAR, fflags, 0, j) != -1);
-
- if (!wp->is_qdir)
+ if (!job_assumes(j, buf != NULL)) {
return;
-
- if (-1 == (qdir_file_cnt = dir_has_files(wp->name))) {
- job_log_error(j, LOG_ERR, "dir_has_files(\"%s\", ...)", wp->name);
- } else if (qdir_file_cnt > 0) {
- job_start(j);
}
-}
-void
-watchpath_callback(struct jobcb *j, struct kevent *kev)
-{
- struct watchpath *wp;
- int dir_file_cnt;
-
- SLIST_FOREACH(wp, &j->vnodes, sle) {
- if (wp->fd == (int)kev->ident)
- break;
- }
+ bufindex = buf;
- launchd_assumes(wp != NULL);
+ rsz = read(j->log_redirect_fd, buf, BIG_PIPE_SIZE);
- if ((NOTE_DELETE|NOTE_RENAME|NOTE_REVOKE) & kev->fflags) {
- job_log(j, LOG_DEBUG, "Path invalidated: %s", wp->name);
- launchd_assumes(close(wp->fd) == 0);
- wp->fd = -1; /* this will get fixed in watchpath_watch() */
- } else if (!wp->is_qdir) {
- job_log(j, LOG_DEBUG, "Watch path modified: %s", wp->name);
+ if (unlikely(rsz == 0)) {
+ job_log(j, LOG_DEBUG, "Standard out/error pipe closed");
+ close_log_redir = true;
+ } else if (rsz == -1) {
+ if (!job_assumes(j, errno == EAGAIN)) {
+ close_log_redir = true;
+ }
} else {
- job_log(j, LOG_DEBUG, "Queue directory modified: %s", wp->name);
+ buf[rsz] = '\0';
- if (-1 == (dir_file_cnt = dir_has_files(wp->name))) {
- job_log_error(j, LOG_ERR, "dir_has_files(\"%s\", ...)", wp->name);
- } else if (0 == dir_file_cnt) {
- job_log(j, LOG_DEBUG, "Spurious wake up, directory is empty again: %s", wp->name);
- return;
+ while ((msg = strsep(&bufindex, "\n\r"))) {
+ if (msg[0]) {
+ job_log_stdouterr2(j, "%s", msg);
+ }
}
}
- job_start(j);
-}
-
-bool
-calendarinterval_new_from_obj(struct jobcb *j, launch_data_t obj)
-{
- launch_data_t tmp_k;
- struct tm tmptm;
-
- memset(&tmptm, 0, sizeof(0));
-
- tmptm.tm_min = -1;
- tmptm.tm_hour = -1;
- tmptm.tm_mday = -1;
- tmptm.tm_wday = -1;
- tmptm.tm_mon = -1;
-
- if (LAUNCH_DATA_DICTIONARY != launch_data_get_type(obj))
- return false;
-
- if ((tmp_k = launch_data_dict_lookup(obj, LAUNCH_JOBKEY_CAL_MINUTE)))
- tmptm.tm_min = launch_data_get_integer(tmp_k);
- if ((tmp_k = launch_data_dict_lookup(obj, LAUNCH_JOBKEY_CAL_HOUR)))
- tmptm.tm_hour = launch_data_get_integer(tmp_k);
- if ((tmp_k = launch_data_dict_lookup(obj, LAUNCH_JOBKEY_CAL_DAY)))
- tmptm.tm_mday = launch_data_get_integer(tmp_k);
- if ((tmp_k = launch_data_dict_lookup(obj, LAUNCH_JOBKEY_CAL_WEEKDAY)))
- tmptm.tm_wday = launch_data_get_integer(tmp_k);
- if ((tmp_k = launch_data_dict_lookup(obj, LAUNCH_JOBKEY_CAL_MONTH)))
- tmptm.tm_mon = launch_data_get_integer(tmp_k);
+ free(buf);
- return calendarinterval_new(j, &tmptm);
+ if (unlikely(close_log_redir)) {
+ (void)job_assumes(j, runtime_close(j->log_redirect_fd) != -1);
+ j->log_redirect_fd = 0;
+ job_dispatch(j, false);
+ }
}
-bool
-calendarinterval_new(struct jobcb *j, struct tm *w)
+void
+job_kill(job_t j)
{
- struct calendarinterval *ci = calloc(1, sizeof(struct calendarinterval));
-
- if (!launchd_assumes(ci != NULL))
- return false;
-
- ci->when = *w;
+ if (unlikely(!j->p || j->anonymous)) {
+ return;
+ }
- SLIST_INSERT_HEAD(&j->cal_intervals, ci, sle);
+ (void)job_assumes(j, runtime_kill(j->p, SIGKILL) != -1);
- calendarinterval_setalarm(j, ci);
+ j->sent_sigkill = true;
+ (void)job_assumes(j, kevent_mod((uintptr_t)&j->exit_timeout, EVFILT_TIMER, EV_ADD|EV_ONESHOT, NOTE_SECONDS, LAUNCHD_SIGKILL_TIMER, j) != -1);
- return true;
+ job_log(j, LOG_DEBUG, "Sent SIGKILL signal");
}
void
-calendarinterval_delete(struct jobcb *j, struct calendarinterval *ci)
+job_open_shutdown_transaction(job_t j)
{
- launchd_assumes(kevent_mod((uintptr_t)ci, EVFILT_TIMER, EV_DELETE, 0, 0, NULL) != -1);
-
- SLIST_REMOVE(&j->cal_intervals, ci, calendarinterval, sle);
-
- free(ci);
+ if (j->kill_via_shmem) {
+ if (j->shmem) {
+ job_log(j, LOG_DEBUG, "Opening shutdown transaction for job.");
+ (void)__sync_add_and_fetch(&j->shmem->vp_shmem_transaction_cnt, 1);
+ } else {
+ job_log(j, LOG_DEBUG, "Job wants to be dirty at shutdown, but it has not set up shared memory. Treating normally.");
+ j->dirty_at_shutdown = false;
+ }
+ } else {
+ job_log(j, LOG_DEBUG, "Job wants to be dirty at shutdown, but it is not Instant Off-compliant. Treating normally.");
+ j->dirty_at_shutdown = false;
+ }
}
void
-calendarinterval_callback(struct jobcb *j, struct kevent *kev)
+job_close_shutdown_transaction(job_t j)
{
- struct calendarinterval *ci;
-
- SLIST_FOREACH(ci, &j->cal_intervals, sle) {
- if ((uintptr_t)ci == kev->ident)
- break;
- }
-
- if (launchd_assumes(ci != NULL)) {
- calendarinterval_setalarm(j, ci);
- job_start(j);
+ if (j->dirty_at_shutdown) {
+ job_log(j, LOG_DEBUG, "Closing shutdown transaction for job.");
+ if (__sync_sub_and_fetch(&j->shmem->vp_shmem_transaction_cnt, 1) == -1) {
+ job_log(j, LOG_DEBUG, "Job is now clean. Killing.");
+ job_kill(j);
+ }
+ j->dirty_at_shutdown = false;
}
}
-bool
-socketgroup_new(struct jobcb *j, const char *name, int *fds, unsigned int fd_cnt, bool junkfds)
+void
+job_log_children_without_exec(job_t j)
{
- struct socketgroup *sg = calloc(1, sizeof(struct socketgroup) + strlen(name) + 1);
-
- if (!launchd_assumes(sg != NULL))
- return false;
-
- sg->fds = calloc(1, fd_cnt * sizeof(int));
- sg->fd_cnt = fd_cnt;
- sg->junkfds = junkfds;
+ pid_t *pids = NULL;
+ size_t len = sizeof(pid_t) * get_kern_max_proc();
+ int i = 0, kp_cnt = 0;
+
+ if (!do_apple_internal_logging || j->anonymous || j->per_user) {
+ return;
+ }
- if (!launchd_assumes(sg->fds != NULL)) {
- free(sg);
- return false;
+ if (!job_assumes(j, (pids = malloc(len)) != NULL)) {
+ return;
+ }
+ if (!job_assumes(j, (kp_cnt = proc_listchildpids(j->p, pids, len)) != -1)) {
+ goto out;
}
- memcpy(sg->fds, fds, fd_cnt * sizeof(int));
- strcpy(sg->name, name);
+ for (i = 0; i < kp_cnt; i++) {
+ struct proc_bsdshortinfo proc;
+ if (proc_pidinfo(pids[i], PROC_PIDT_SHORTBSDINFO, 1, &proc, PROC_PIDT_SHORTBSDINFO_SIZE) == 0) {
+ if (errno != ESRCH) {
+ job_assumes(j, errno == 0);
+ }
+ continue;
+ }
+ if (proc.pbsi_flags & P_EXEC) {
+ continue;
+ }
- SLIST_INSERT_HEAD(&j->sockets, sg, sle);
+ job_log(j, LOG_DEBUG, "Called *fork(). Please switch to posix_spawn*(), pthreads or launchd. Child PID %u", pids[i]);
+ }
- return true;
+out:
+ free(pids);
}
void
-socketgroup_delete(struct jobcb *j, struct socketgroup *sg)
+job_cleanup_after_tracer(job_t j)
{
- unsigned int i;
-
- for (i = 0; i < sg->fd_cnt; i++)
- launchd_assumes(close(sg->fds[i]) != -1);
-
- SLIST_REMOVE(&j->sockets, sg, socketgroup, sle);
-
- free(sg->fds);
- free(sg);
+ j->tracing_pid = 0;
+ if (j->reap_after_trace) {
+ job_log(j, LOG_DEBUG | LOG_CONSOLE, "Reaping job now that attached tracer is gone.");
+ struct kevent kev;
+ EV_SET(&kev, j->p, 0, 0, NOTE_EXIT, 0, 0);
+
+ /* Fake a kevent to keep our logic consistent. */
+ job_callback_proc(j, &kev);
+
+ /* Normally, after getting a EVFILT_PROC event, we do garbage collection
+ * on the root job manager. To make our fakery complete, we will do garbage
+ * collection at the beginning of the next run loop cycle (after we're done
+ * draining the current queue of kevents).
+ */
+ (void)job_assumes(j, kevent_mod((uintptr_t)&root_jobmgr->reboot_flags, EVFILT_TIMER, EV_ADD | EV_ONESHOT, NOTE_NSECONDS, 1, root_jobmgr) != -1);
+ }
}
void
-socketgroup_ignore(struct jobcb *j, struct socketgroup *sg)
+job_callback_proc(job_t j, struct kevent *kev)
{
- char buf[10000];
- unsigned int i, buf_off = 0;
-
- if (sg->junkfds)
- return;
-
- for (i = 0; i < sg->fd_cnt; i++)
- buf_off += sprintf(buf + buf_off, " %d", sg->fds[i]);
-
- job_log(j, LOG_DEBUG, "Ignoring Sockets:%s", buf);
+ bool program_changed = false;
+ int fflags = kev->fflags;
+
+ job_log(j, LOG_DEBUG, "EVFILT_PROC event for job.");
+ log_kevent_struct(LOG_DEBUG, kev, 0);
+
+ if (fflags & NOTE_EXIT) {
+ if (j->p == (pid_t)kev->ident && !j->anonymous) {
+ /* Note that the third argument to proc_pidinfo() is a magic argument for
+ * PROC_PIDT_SHORTBSDINFO. Specifically, passing 1 means "don't fail on a zombie
+ * PID".
+ */
+ struct proc_bsdshortinfo proc;
+ if (job_assumes(j, proc_pidinfo(j->p, PROC_PIDT_SHORTBSDINFO, 1, &proc, PROC_PIDT_SHORTBSDINFO_SIZE) > 0)) {
+ if (!job_assumes(j, (pid_t)proc.pbsi_ppid == getpid())) {
+ /* Someone has attached to the process with ptrace(). There's a race here.
+ * If we determine that we are not the parent process and then fail to attach
+ * a kevent to the parent PID (who is probably using ptrace()), we can take that as an
+ * indication that the parent exited between sysctl(3) and kevent_mod(). The
+ * reparenting of the PID should be atomic to us, so in that case, we reap the
+ * job as normal.
+ *
+ * Otherwise, we wait for the death of the parent tracer and then reap, just as we
+ * would if a job died while we were sampling it at shutdown.
+ *
+ * Note that we foolishly assume that in the process *tree* a node cannot be its
+ * own parent. Apparently, that is not correct. If this is the case, we forsake
+ * the process to its own devices. Let it reap itself.
+ */
+ if (!job_assumes(j, proc.pbsi_ppid != kev->ident)) {
+ job_log(j, LOG_WARNING, "Job is its own parent and has (somehow) exited. Leaving it to waste away.");
+ return;
+ }
+ if (job_assumes(j, kevent_mod(proc.pbsi_ppid, EVFILT_PROC, EV_ADD, NOTE_EXIT, 0, j) != -1)) {
+ j->tracing_pid = proc.pbsi_ppid;
+ j->reap_after_trace = true;
+ return;
+ }
+ }
+ }
+ } else if (!j->anonymous) {
+ if (j->tracing_pid == (pid_t)kev->ident) {
+ job_cleanup_after_tracer(j);
+
+ return;
+ } else if (j->tracing_pid && !j->reap_after_trace) {
+ /* The job exited before our sample completed. */
+ job_log(j, LOG_DEBUG | LOG_CONSOLE, "Job has exited. Will reap after tracing PID %i exits.", j->tracing_pid);
+ j->reap_after_trace = true;
+ return;
+ }
+ }
+ }
+
+ if (fflags & NOTE_EXEC) {
+ program_changed = true;
+
+ if (j->anonymous) {
+ struct proc_bsdshortinfo proc;
+ if (proc_pidinfo(j->p, PROC_PIDT_SHORTBSDINFO, 1, &proc, PROC_PIDT_SHORTBSDINFO_SIZE) > 0) {
+ char newlabel[1000];
+
+ snprintf(newlabel, sizeof(newlabel), "%p.anonymous.%s", j, proc.pbsi_comm);
+
+ job_log(j, LOG_INFO, "Program changed. Updating the label to: %s", newlabel);
+ j->lastlookup = NULL;
+ j->lastlookup_gennum = 0;
+
+ LIST_REMOVE(j, label_hash_sle);
+ strcpy((char *)j->label, newlabel);
+
+ jobmgr_t where2put = root_jobmgr;
+ if (j->mgr->properties & BOOTSTRAP_PROPERTY_XPC_DOMAIN) {
+ where2put = j->mgr;
+ }
+ LIST_INSERT_HEAD(&where2put->label_hash[hash_label(j->label)], j, label_hash_sle);
+ } else if (errno != ESRCH) {
+ job_assumes(j, errno == 0);
+ }
+ } else {
+ if (j->spawn_reply_port) {
+ errno = job_mig_spawn2_reply(j->spawn_reply_port, BOOTSTRAP_SUCCESS, j->p, j->exit_status_port);
+ if (errno) {
+ if (errno != MACH_SEND_INVALID_DEST) {
+ (void)job_assumes(j, errno == KERN_SUCCESS);
+ }
+ (void)job_assumes(j, launchd_mport_close_recv(j->exit_status_port) == KERN_SUCCESS);
+ }
- for (i = 0; i < sg->fd_cnt; i++)
- launchd_assumes(kevent_mod(sg->fds[i], EVFILT_READ, EV_DELETE, 0, 0, NULL) != -1);
-}
+ j->spawn_reply_port = MACH_PORT_NULL;
+ j->exit_status_port = MACH_PORT_NULL;
+ }
-void
-socketgroup_watch(struct jobcb *j, struct socketgroup *sg)
-{
- char buf[10000];
- unsigned int i, buf_off = 0;
+ if (j->xpc_service && j->did_exec) {
+ j->xpcproxy_did_exec = true;
+ }
- if (sg->junkfds)
- return;
+ j->did_exec = true;
+ job_log(j, LOG_DEBUG, "Program changed");
+ }
+ }
- for (i = 0; i < sg->fd_cnt; i++)
- buf_off += sprintf(buf + buf_off, " %d", sg->fds[i]);
+ if (fflags & NOTE_FORK) {
+ job_log(j, LOG_DEBUG, "fork()ed%s", program_changed ? ". For this message only: We don't know whether this event happened before or after execve()." : "");
+ job_log_children_without_exec(j);
+ }
- job_log(j, LOG_DEBUG, "Watching sockets:%s", buf);
+ if (fflags & NOTE_EXIT) {
+ job_reap(j);
- for (i = 0; i < sg->fd_cnt; i++)
- launchd_assumes(kevent_mod(sg->fds[i], EVFILT_READ, EV_ADD, 0, 0, j) != -1);
+ if (j->anonymous) {
+ job_remove(j);
+ j = NULL;
+ } else {
+ j = job_dispatch(j, false);
+ }
+ }
}
void
-socketgroup_callback(struct jobcb *j, struct kevent *kev)
+job_callback_timer(job_t j, void *ident)
{
- job_start(j);
-}
+ if (j == ident) {
+ job_log(j, LOG_DEBUG, "j == ident (%p)", ident);
+ job_dispatch(j, true);
+ } else if (&j->semaphores == ident) {
+ job_log(j, LOG_DEBUG, "&j->semaphores == ident (%p)", ident);
+ job_dispatch(j, false);
+ } else if (&j->start_interval == ident) {
+ job_log(j, LOG_DEBUG, "&j->start_interval == ident (%p)", ident);
+ j->start_pending = true;
+ job_dispatch(j, false);
+ } else if (&j->exit_timeout == ident) {
+ if (!job_assumes(j, j->p != 0)) {
+ return;
+ }
-bool
-envitem_new(struct jobcb *j, const char *k, const char *v, bool global)
-{
- struct envitem *ei = calloc(1, sizeof(struct envitem) + strlen(k) + 1 + strlen(v) + 1);
+ if (j->sent_sigkill) {
+ uint64_t td = runtime_get_nanoseconds_since(j->sent_signal_time);
- if (!launchd_assumes(ei != NULL))
- return false;
+ td /= NSEC_PER_SEC;
+ td -= j->clean_kill ? 0 : j->exit_timeout;
- strcpy(ei->key, k);
- ei->value = ei->key + strlen(k) + 1;
- strcpy(ei->value, v);
+ job_log(j, LOG_WARNING | LOG_CONSOLE, "Job has not died after being %skilled %llu seconds ago. Simulating exit.", j->clean_kill ? "cleanly " : "", td);
+ j->workaround9359725 = true;
- if (global) {
- SLIST_INSERT_HEAD(&j->global_env, ei, sle);
+ if (g_trap_sigkill_bugs) {
+ job_log(j, LOG_NOTICE | LOG_CONSOLE, "Trapping into kernel debugger. You can continue the machine after it has been debugged, and shutdown will proceed normally.");
+ (void)job_assumes(j, host_reboot(mach_host_self(), HOST_REBOOT_DEBUGGER) == KERN_SUCCESS);
+ }
+
+ struct kevent bogus_exit;
+ EV_SET(&bogus_exit, j->p, EVFILT_PROC, 0, NOTE_EXIT, 0, 0);
+ jobmgr_callback(j->mgr, &bogus_exit);
+ } else {
+ if (unlikely(j->debug_before_kill)) {
+ job_log(j, LOG_NOTICE, "Exit timeout elapsed. Entering the kernel debugger");
+ (void)job_assumes(j, host_reboot(mach_host_self(), HOST_REBOOT_DEBUGGER) == KERN_SUCCESS);
+ }
+
+ job_log(j, LOG_WARNING | LOG_CONSOLE, "Exit timeout elapsed (%u seconds). Killing", j->exit_timeout);
+ job_kill(j);
+ }
} else {
- SLIST_INSERT_HEAD(&j->env, ei, sle);
+ (void)job_assumes(j, false);
}
-
- return true;
}
void
-envitem_delete(struct jobcb *j, struct envitem *ei, bool global)
+job_callback_read(job_t j, int ident)
{
- if (global) {
- SLIST_REMOVE(&j->global_env, ei, envitem, sle);
+ if (ident == j->log_redirect_fd) {
+ job_log_stdouterr(j);
+ } else if (ident == j->stdin_fd) {
+ job_dispatch(j, true);
} else {
- SLIST_REMOVE(&j->env, ei, envitem, sle);
+ socketgroup_callback(j);
}
-
- free(ei);
}
void
-envitem_setup(launch_data_t obj, const char *key, void *context)
+jobmgr_reap_bulk(jobmgr_t jm, struct kevent *kev)
{
- struct jobcb *j = context;
+ jobmgr_t jmi;
+ job_t j;
- if (launch_data_get_type(obj) != LAUNCH_DATA_STRING)
- return;
+ SLIST_FOREACH(jmi, &jm->submgrs, sle) {
+ jobmgr_reap_bulk(jmi, kev);
+ }
- envitem_new(j, key, launch_data_get_string(obj), j->importing_global_env);
+ if ((j = jobmgr_find_by_pid(jm, (pid_t)kev->ident, false))) {
+ kev->udata = j;
+ job_callback(j, kev);
+ }
}
-bool
-limititem_update(struct jobcb *j, int w, rlim_t r)
+void
+jobmgr_callback(void *obj, struct kevent *kev)
{
- struct limititem *li;
-
- SLIST_FOREACH(li, &j->limits, sle) {
- if (li->which == w)
- break;
- }
-
- if (li == NULL) {
- li = calloc(1, sizeof(struct limititem));
-
- if (!launchd_assumes(li != NULL))
- return false;
+ jobmgr_t jm = obj;
+ job_t ji;
- li->which = w;
- }
+ switch (kev->filter) {
+ case EVFILT_PROC:
+ jobmgr_reap_bulk(jm, kev);
+ root_jobmgr = jobmgr_do_garbage_collection(root_jobmgr);
+ break;
+ case EVFILT_SIGNAL:
+ switch (kev->ident) {
+ case SIGTERM:
+ jobmgr_log(jm, LOG_DEBUG, "Got SIGTERM. Shutting down.");
+ return launchd_shutdown();
+ case SIGUSR1:
+ return calendarinterval_callback();
+ case SIGUSR2:
+ fake_shutdown_in_progress = true;
+ runtime_setlogmask(LOG_UPTO(LOG_DEBUG));
+
+ runtime_closelog(); /* HACK -- force 'start' time to be set */
+
+ if (pid1_magic) {
+ int64_t now = runtime_get_wall_time();
+
+ jobmgr_log(jm, LOG_NOTICE, "Anticipatory shutdown began at: %lld.%06llu", now / USEC_PER_SEC, now % USEC_PER_SEC);
+
+ LIST_FOREACH(ji, &root_jobmgr->jobs, sle) {
+ if (ji->per_user && ji->p) {
+ (void)job_assumes(ji, runtime_kill(ji->p, SIGUSR2) != -1);
+ }
+ }
+ } else {
+ jobmgr_log(jm, LOG_NOTICE, "Anticipatory per-user launchd shutdown");
+ }
- if (j->importing_hard_limits) {
- li->lim.rlim_max = r;
- li->sethard = true;
- } else {
- li->lim.rlim_cur = r;
- li->setsoft = true;
+ return;
+ default:
+ return (void)jobmgr_assumes(jm, false);
+ }
+ break;
+ case EVFILT_FS:
+ if (kev->fflags & VQ_MOUNT) {
+ jobmgr_dispatch_all(jm, true);
+ }
+ jobmgr_dispatch_all_semaphores(jm);
+ break;
+ case EVFILT_TIMER:
+ if (kev->ident == (uintptr_t)&sorted_calendar_events) {
+ calendarinterval_callback();
+ } else if (kev->ident == (uintptr_t)jm) {
+ jobmgr_log(jm, LOG_DEBUG, "Shutdown timer firing.");
+ jobmgr_still_alive_with_check(jm);
+ } else if (kev->ident == (uintptr_t)&jm->reboot_flags) {
+ jobmgr_do_garbage_collection(jm);
+ } else if (kev->ident == (uintptr_t)&g_runtime_busy_time) {
+ jobmgr_log(jm, LOG_DEBUG, "Idle exit timer fired. Shutting down.");
+ if (jobmgr_assumes(jm, runtime_busy_cnt == 0)) {
+ return launchd_shutdown();
+ }
+ }
+ break;
+ case EVFILT_VNODE:
+ if (kev->ident == (uintptr_t)s_no_hang_fd) {
+ int _no_hang_fd = open("/dev/autofs_nowait", O_EVTONLY | O_NONBLOCK);
+ if (unlikely(_no_hang_fd != -1)) {
+ jobmgr_log(root_jobmgr, LOG_DEBUG, "/dev/autofs_nowait has appeared!");
+ (void)jobmgr_assumes(root_jobmgr, kevent_mod((uintptr_t)s_no_hang_fd, EVFILT_VNODE, EV_DELETE, 0, 0, NULL) != -1);
+ (void)jobmgr_assumes(root_jobmgr, runtime_close(s_no_hang_fd) != -1);
+ s_no_hang_fd = _fd(_no_hang_fd);
+ }
+ } else if (pid1_magic && g_console && kev->ident == (uintptr_t)fileno(g_console)) {
+ int cfd = -1;
+ if (launchd_assumes((cfd = open(_PATH_CONSOLE, O_WRONLY | O_NOCTTY)) != -1)) {
+ _fd(cfd);
+ if (!launchd_assumes((g_console = fdopen(cfd, "w")) != NULL)) {
+ close(cfd);
+ }
+ }
+ }
+ break;
+ default:
+ return (void)jobmgr_assumes(jm, false);
}
-
- return true;
}
void
-limititem_delete(struct jobcb *j, struct limititem *li)
+job_callback(void *obj, struct kevent *kev)
{
- SLIST_REMOVE(&j->limits, li, limititem, sle);
+ job_t j = obj;
- free(li);
+ job_log(j, LOG_DEBUG, "Dispatching kevent callback.");
+
+ switch (kev->filter) {
+ case EVFILT_PROC:
+ return job_callback_proc(j, kev);
+ case EVFILT_TIMER:
+ return job_callback_timer(j, (void *) kev->ident);
+ case EVFILT_VNODE:
+ return semaphoreitem_callback(j, kev);
+ case EVFILT_READ:
+ return job_callback_read(j, (int) kev->ident);
+ case EVFILT_MACHPORT:
+ return (void)job_dispatch(j, true);
+ default:
+ return (void)job_assumes(j, false);
+ }
}
void
-limititem_setup(launch_data_t obj, const char *key, void *context)
+job_start(job_t j)
{
- struct jobcb *j = context;
- int i, limits_cnt = (sizeof(launchd_keys2limits) / sizeof(launchd_keys2limits[0]));
- rlim_t rl;
-
- if (launch_data_get_type(obj) != LAUNCH_DATA_INTEGER)
+ uint64_t td;
+ int spair[2];
+ int execspair[2];
+ int oepair[2];
+ char nbuf[64];
+ pid_t c;
+ bool sipc = false;
+ u_int proc_fflags = NOTE_EXIT|NOTE_FORK|NOTE_EXEC;
+
+ if (!job_assumes(j, j->mgr != NULL)) {
return;
-
- rl = launch_data_get_integer(obj);
-
- for (i = 0; i < limits_cnt; i++) {
- if (strcasecmp(launchd_keys2limits[i].key, key) == 0)
- break;
}
-
- if (i == limits_cnt)
+
+ if (unlikely(job_active(j))) {
+ job_log(j, LOG_DEBUG, "Already started");
return;
+ }
+
+ /*
+ * Some users adjust the wall-clock and then expect software to not notice.
+ * Therefore, launchd must use an absolute clock instead of the wall clock
+ * wherever possible.
+ */
+ td = runtime_get_nanoseconds_since(j->start_time);
+ td /= NSEC_PER_SEC;
+
+ if (j->start_time && (td < j->min_run_time) && !j->legacy_mach_job && !j->inetcompat) {
+ time_t respawn_delta = j->min_run_time - (uint32_t)td;
+
+ /*
+ * We technically should ref-count throttled jobs to prevent idle exit,
+ * but we're not directly tracking the 'throttled' state at the moment.
+ */
+ int level = LOG_WARNING;
+ if (!j->did_exec && ((j->fail_cnt - 1) % LAUNCHD_LOG_FAILED_EXEC_FREQ) != 0) {
+ level = LOG_DEBUG;
+ }
- limititem_update(j, launchd_keys2limits[i].val, rl);
-}
-
-bool
-job_useless(struct jobcb *j)
-{
- if (j->unload_at_exit) {
- job_log(j, LOG_INFO, "Exited. Was only configured to run once.");
- return true;
- } else if (shutdown_in_progress) {
- job_log(j, LOG_INFO, "Exited while shutdown in progress.");
- return true;
- } else if (j->failed_exits >= LAUNCHD_FAILED_EXITS_THRESHOLD) {
- job_log(j, LOG_WARNING, "too many failures in succession");
- return true;
- } else if (!j->checkedin && (!SLIST_EMPTY(&j->sockets) || !SLIST_EMPTY(&j->machservices))) {
- job_log(j, LOG_WARNING, "Failed to check-in!");
- return true;
- } else if (j->legacy_mach_job && SLIST_EMPTY(&j->machservices)) {
- job_log(j, LOG_INFO, "Garbage collecting");
- return true;
+ job_log(j, level, "Throttling respawn: Will start in %ld seconds", respawn_delta);
+ (void)job_assumes(j, kevent_mod((uintptr_t)j, EVFILT_TIMER, EV_ADD|EV_ONESHOT, NOTE_SECONDS, respawn_delta, j) != -1);
+ job_ignore(j);
+ return;
+ }
+
+ if (likely(!j->legacy_mach_job)) {
+ sipc = ((!SLIST_EMPTY(&j->sockets) || !SLIST_EMPTY(&j->machservices)) && !j->deny_job_creation) || j->embedded_special_privileges;
}
- return false;
+ if (sipc) {
+ (void)job_assumes(j, socketpair(AF_UNIX, SOCK_STREAM, 0, spair) != -1);
+ }
+
+ (void)job_assumes(j, socketpair(AF_UNIX, SOCK_STREAM, 0, execspair) != -1);
+
+ if (likely(!j->legacy_mach_job) && job_assumes(j, pipe(oepair) != -1)) {
+ j->log_redirect_fd = _fd(oepair[0]);
+ (void)job_assumes(j, fcntl(j->log_redirect_fd, F_SETFL, O_NONBLOCK) != -1);
+ (void)job_assumes(j, kevent_mod(j->log_redirect_fd, EVFILT_READ, EV_ADD, 0, 0, j) != -1);
+ }
+
+ switch (c = runtime_fork(j->weird_bootstrap ? j->j_port : j->mgr->jm_port)) {
+ case -1:
+ job_log_error(j, LOG_ERR, "fork() failed, will try again in one second");
+ (void)job_assumes(j, kevent_mod((uintptr_t)j, EVFILT_TIMER, EV_ADD|EV_ONESHOT, NOTE_SECONDS, 1, j) != -1);
+ job_ignore(j);
+
+ (void)job_assumes(j, runtime_close(execspair[0]) == 0);
+ (void)job_assumes(j, runtime_close(execspair[1]) == 0);
+ if (sipc) {
+ (void)job_assumes(j, runtime_close(spair[0]) == 0);
+ (void)job_assumes(j, runtime_close(spair[1]) == 0);
+ }
+ if (likely(!j->legacy_mach_job)) {
+ (void)job_assumes(j, runtime_close(oepair[0]) != -1);
+ (void)job_assumes(j, runtime_close(oepair[1]) != -1);
+ j->log_redirect_fd = 0;
+ }
+ break;
+ case 0:
+ if (unlikely(_vproc_post_fork_ping())) {
+ _exit(EXIT_FAILURE);
+ }
+ if (!j->legacy_mach_job) {
+ (void)job_assumes(j, dup2(oepair[1], STDOUT_FILENO) != -1);
+ (void)job_assumes(j, dup2(oepair[1], STDERR_FILENO) != -1);
+ (void)job_assumes(j, runtime_close(oepair[1]) != -1);
+ }
+ (void)job_assumes(j, runtime_close(execspair[0]) == 0);
+ /* wait for our parent to say they've attached a kevent to us */
+ read(_fd(execspair[1]), &c, sizeof(c));
+
+ if (sipc) {
+ (void)job_assumes(j, runtime_close(spair[0]) == 0);
+ snprintf(nbuf, sizeof(nbuf), "%d", spair[1]);
+ setenv(LAUNCHD_TRUSTED_FD_ENV, nbuf, 1);
+ }
+ job_start_child(j);
+ break;
+ default:
+ j->start_time = runtime_get_opaque_time();
+
+ job_log(j, LOG_DEBUG, "Started as PID: %u", c);
+
+ j->did_exec = false;
+ j->xpcproxy_did_exec = false;
+ j->checkedin = false;
+ j->start_pending = false;
+ j->reaped = false;
+ j->crashed = false;
+ j->stopped = false;
+ if (j->needs_kickoff) {
+ j->needs_kickoff = false;
+
+ if (SLIST_EMPTY(&j->semaphores)) {
+ j->ondemand = false;
+ }
+ }
+
+ if (j->has_console) {
+ g_wsp = c;
+ }
+
+ runtime_add_ref();
+ total_children++;
+ LIST_INSERT_HEAD(&j->mgr->active_jobs[ACTIVE_JOB_HASH(c)], j, pid_hash_sle);
+
+ if (likely(!j->legacy_mach_job)) {
+ (void)job_assumes(j, runtime_close(oepair[1]) != -1);
+ }
+ j->p = c;
+
+ j->mgr->normal_active_cnt++;
+ j->fork_fd = _fd(execspair[0]);
+ (void)job_assumes(j, runtime_close(execspair[1]) == 0);
+ if (sipc) {
+ (void)job_assumes(j, runtime_close(spair[1]) == 0);
+ ipc_open(_fd(spair[0]), j);
+ }
+ if (job_assumes(j, kevent_mod(c, EVFILT_PROC, EV_ADD, proc_fflags, 0, root_jobmgr ? root_jobmgr : j->mgr) != -1)) {
+ job_ignore(j);
+ } else {
+ job_reap(j);
+ }
+
+ j->wait4debugger_oneshot = false;
+
+ struct envitem *ei = NULL, *et = NULL;
+ SLIST_FOREACH_SAFE(ei, &j->env, sle, et) {
+ if (ei->one_shot) {
+ SLIST_REMOVE(&j->env, ei, envitem, sle);
+ }
+ }
+
+ if (likely(!j->stall_before_exec)) {
+ job_uncork_fork(j);
+ }
+ break;
+ }
}
-bool
-job_keepalive(struct jobcb *j)
+void
+job_start_child(job_t j)
{
- mach_msg_type_number_t statusCnt;
- mach_port_status_t status;
- struct semaphoreitem *si;
- struct machservice *ms;
- struct stat sb;
- bool good_exit = (WIFEXITED(j->last_exit_status) && WEXITSTATUS(j->last_exit_status) == 0);
- bool dispatch_others = false;
+ typeof(posix_spawn) *psf;
+ const char *file2exec = "/usr/libexec/launchproxy";
+ const char **argv;
+ posix_spawnattr_t spattr;
+ int gflags = GLOB_NOSORT|GLOB_NOCHECK|GLOB_TILDE|GLOB_DOOFFS;
+ glob_t g;
+ short spflags = POSIX_SPAWN_SETEXEC;
+ size_t binpref_out_cnt = 0;
+ size_t i;
- if (j->runatload && j->start_time == 0) {
- job_log(j, LOG_DEBUG, "KeepAlive check: job needs to run at least once.");
- return true;
+ (void)job_assumes(j, posix_spawnattr_init(&spattr) == 0);
+
+ job_setup_attributes(j);
+
+ if (unlikely(j->argv && j->globargv)) {
+ g.gl_offs = 1;
+ for (i = 0; i < j->argc; i++) {
+ if (i > 0) {
+ gflags |= GLOB_APPEND;
+ }
+ if (glob(j->argv[i], gflags, NULL, &g) != 0) {
+ job_log_error(j, LOG_ERR, "glob(\"%s\")", j->argv[i]);
+ exit(EXIT_FAILURE);
+ }
+ }
+ g.gl_pathv[0] = (char *)file2exec;
+ argv = (const char **)g.gl_pathv;
+ } else if (likely(j->argv)) {
+ argv = alloca((j->argc + 2) * sizeof(char *));
+ argv[0] = file2exec;
+ for (i = 0; i < j->argc; i++) {
+ argv[i + 1] = j->argv[i];
+ }
+ argv[i + 1] = NULL;
+ } else {
+ argv = alloca(3 * sizeof(char *));
+ argv[0] = file2exec;
+ argv[1] = j->prog;
+ argv[2] = NULL;
}
- if (!j->ondemand) {
- job_log(j, LOG_DEBUG, "KeepAlive check: job configured to run continuously.");
- return true;
+ if (likely(!j->inetcompat)) {
+ argv++;
}
- SLIST_FOREACH(ms, &j->machservices, sle) {
- statusCnt = MACH_PORT_RECEIVE_STATUS_COUNT;
- if (mach_port_get_attributes(mach_task_self(), ms->port, MACH_PORT_RECEIVE_STATUS,
- (mach_port_info_t)&status, &statusCnt) != KERN_SUCCESS)
- continue;
- if (status.mps_msgcount) {
- job_log(j, LOG_DEBUG, "KeepAlive check: job restarted due to %d queued Mach messages on service: %s",
- status.mps_msgcount, ms->name);
- return true;
+ if (unlikely(j->wait4debugger || j->wait4debugger_oneshot)) {
+ if (!j->legacy_LS_job) {
+ job_log(j, LOG_WARNING, "Spawned and waiting for the debugger to attach before continuing...");
}
+ spflags |= POSIX_SPAWN_START_SUSPENDED;
}
+ if (unlikely(j->disable_aslr)) {
+ spflags |= _POSIX_SPAWN_DISABLE_ASLR;
+ }
+ spflags |= j->pstype;
- SLIST_FOREACH(si, &j->semaphores, sle) {
- bool wanted_state = false;
- switch (si->why) {
- case NETWORK_UP:
- wanted_state = true;
- case NETWORK_DOWN:
- if (network_up == wanted_state) {
- job_log(j, LOG_DEBUG, "KeepAlive check: job configured to run while the network is %s.",
- wanted_state ? "up" : "down");
- return true;
- }
- break;
- case SUCCESSFUL_EXIT:
- wanted_state = true;
- case FAILED_EXIT:
- if (good_exit == wanted_state) {
- job_log(j, LOG_DEBUG, "KeepAlive check: job configured to run while the exit state was %s.",
- wanted_state ? "successful" : "failure");
- return true;
+ (void)job_assumes(j, posix_spawnattr_setflags(&spattr, spflags) == 0);
+
+ if (unlikely(j->j_binpref_cnt)) {
+ (void)job_assumes(j, posix_spawnattr_setbinpref_np(&spattr, j->j_binpref_cnt, j->j_binpref, &binpref_out_cnt) == 0);
+ (void)job_assumes(j, binpref_out_cnt == j->j_binpref_cnt);
+ }
+
+#if HAVE_QUARANTINE
+ if (j->quarantine_data) {
+ qtn_proc_t qp;
+
+ if (job_assumes(j, qp = qtn_proc_alloc())) {
+ if (job_assumes(j, qtn_proc_init_with_data(qp, j->quarantine_data, j->quarantine_data_sz) == 0)) {
+ (void)job_assumes(j, qtn_proc_apply_to_self(qp) == 0);
}
- break;
- case PATH_EXISTS:
- wanted_state = true;
- case PATH_MISSING:
- if ((bool)(stat(si->what, &sb) == 0) == wanted_state) {
- job_log(j, LOG_DEBUG, "KeepAlive check: job configured to run while the following path %s: %s",
- wanted_state ? "exists" : "is missing", si->what);
- return true;
+ }
+ }
+#endif
+
+#if HAVE_SANDBOX
+ if (j->seatbelt_profile) {
+ char *seatbelt_err_buf = NULL;
+
+ if (!job_assumes(j, sandbox_init(j->seatbelt_profile, j->seatbelt_flags, &seatbelt_err_buf) != -1)) {
+ if (seatbelt_err_buf) {
+ job_log(j, LOG_ERR, "Sandbox failed to init: %s", seatbelt_err_buf);
}
- dispatch_others = true;
- break;
+ goto out_bad;
}
}
+#endif
- /* Maybe another job has the inverse path based semaphore as this job */
- if (dispatch_others)
- job_dispatch_all_other_semaphores(root_job, j);
+ psf = j->prog ? posix_spawn : posix_spawnp;
- return false;
-}
+ if (likely(!j->inetcompat)) {
+ file2exec = j->prog ? j->prog : argv[0];
+ }
-const char *
-job_prog(struct jobcb *j)
-{
- if (j->prog) {
- return j->prog;
- } else if (j->argv) {
- return j->argv[0];
- } else {
- return "";
+ errno = psf(NULL, file2exec, NULL, &spattr, (char *const *)argv, environ);
+ if (errno != EBADARCH) {
+ int level = LOG_ERR;
+ if ((j->fail_cnt++ % LAUNCHD_LOG_FAILED_EXEC_FREQ) != 0) {
+ level = LOG_DEBUG;
+ }
+ job_log_error(j, level, "posix_spawn(\"%s\", ...)", file2exec);
+ errno = EXIT_FAILURE;
}
+
+#if HAVE_SANDBOX
+out_bad:
+#endif
+ _exit(errno);
}
-bool
-job_active(struct jobcb *j)
+void
+jobmgr_export_env_from_other_jobs(jobmgr_t jm, launch_data_t dict)
{
- struct machservice *ms;
-
- if (j->req_port)
- return true;
+ launch_data_t tmp;
+ struct envitem *ei;
+ job_t ji;
- if (j->p)
- return true;
+ if (jm->parentmgr) {
+ jobmgr_export_env_from_other_jobs(jm->parentmgr, dict);
+ } else {
+ char **tmpenviron = environ;
+ for (; *tmpenviron; tmpenviron++) {
+ char envkey[1024];
+ launch_data_t s = launch_data_alloc(LAUNCH_DATA_STRING);
+ launch_data_set_string(s, strchr(*tmpenviron, '=') + 1);
+ strncpy(envkey, *tmpenviron, sizeof(envkey));
+ *(strchr(envkey, '=')) = '\0';
+ launch_data_dict_insert(dict, s, envkey);
+ }
+ }
- if (j->priv_port_has_senders) {
- if (j->start_time && !j->checkedin) {
- if (j->legacy_mach_job) {
- job_log(j, LOG_NOTICE, "Daemonized. Extremely expensive no-op.");
- } else if (!j->unload_at_exit) {
- job_log(j, LOG_ERR, "Daemonization is not supported under launchd.");
- return false;
+ LIST_FOREACH(ji, &jm->jobs, sle) {
+ SLIST_FOREACH(ei, &ji->global_env, sle) {
+ if ((tmp = launch_data_new_string(ei->value))) {
+ launch_data_dict_insert(dict, tmp, ei->key);
}
}
- return true;
}
+}
- SLIST_FOREACH(ms, &j->machservices, sle) {
- if (ms->isActive)
- return true;
+void
+jobmgr_setup_env_from_other_jobs(jobmgr_t jm)
+{
+ struct envitem *ei;
+ job_t ji;
+
+ if (jm->parentmgr) {
+ jobmgr_setup_env_from_other_jobs(jm->parentmgr);
}
- return false;
+ LIST_FOREACH(ji, &jm->global_env_jobs, global_env_sle) {
+ SLIST_FOREACH(ei, &ji->global_env, sle) {
+ setenv(ei->key, ei->value, 1);
+ }
+ }
}
-pid_t
-launchd_fork(void)
+void
+job_log_pids_with_weird_uids(job_t j)
{
- return job_fork(root_job);
-}
+ size_t len = sizeof(pid_t) * get_kern_max_proc();
+ pid_t *pids = NULL;
+ uid_t u = j->mach_uid;
+ int i = 0, kp_cnt = 0;
+
+ if (!do_apple_internal_logging) {
+ return;
+ }
-pid_t
-job_fork(struct jobcb *j)
-{
- mach_port_t p = j->bs_port;
- pid_t r = -1;
+ pids = malloc(len);
+ if (!job_assumes(j, pids != NULL)) {
+ return;
+ }
- sigprocmask(SIG_BLOCK, &blocked_signals, NULL);
+ runtime_ktrace(RTKT_LAUNCHD_FINDING_WEIRD_UIDS, j->p, u, 0);
+
+ /* libproc actually has some serious performance drawbacks when used over sysctl(3) in
+ * scenarios like this. Whereas sysctl(3) can give us back all the kinfo_proc's in
+ * one kernel call, libproc requires that we get a list of PIDs we're interested in
+ * (in this case, all PIDs on the system) and then get a single proc_bsdshortinfo
+ * struct back in a single call for each one.
+ *
+ * This kind of thing is also more inherently racy than sysctl(3). While sysctl(3)
+ * returns a snapshot, it returns the whole shebang at once. Any PIDs given to us by
+ * libproc could go stale before we call proc_pidinfo().
+ *
+ * Note that proc_list*() APIs return the number of PIDs given back, not the number
+ * of bytes written to the buffer.
+ */
+ if (!job_assumes(j, (kp_cnt = proc_listallpids(pids, len)) != -1)) {
+ goto out;
+ }
- launchd_assumes(launchd_mport_make_send(p) == KERN_SUCCESS);
- launchd_assumes(launchd_set_bport(p) == KERN_SUCCESS);
- launchd_assumes(launchd_mport_deallocate(p) == KERN_SUCCESS);
+ for (i = 0; i < kp_cnt; i++) {
+ struct proc_bsdshortinfo proc;
+ /* We perhaps should not log a bug here if we get ESRCH back, due to the race
+ * detailed above.
+ */
+ if (proc_pidinfo(pids[i], PROC_PIDT_SHORTBSDINFO, 1, &proc, PROC_PIDT_SHORTBSDINFO_SIZE) == 0) {
+ if (errno != ESRCH) {
+ job_assumes(j, errno == 0);
+ }
+ continue;
+ }
+
+ uid_t i_euid = proc.pbsi_uid;
+ uid_t i_uid = proc.pbsi_ruid;
+ uid_t i_svuid = proc.pbsi_svuid;
+ pid_t i_pid = pids[i];
- r = fork();
+ if (i_euid != u && i_uid != u && i_svuid != u) {
+ continue;
+ }
- if (r != 0) {
- launchd_assumes(launchd_set_bport(MACH_PORT_NULL) == KERN_SUCCESS);
- } else if (r == 0) {
- size_t i;
+ job_log(j, LOG_ERR, "PID %u \"%s\" has no account to back it! Real/effective/saved UIDs: %u/%u/%u", i_pid, proc.pbsi_comm, i_uid, i_euid, i_svuid);
- for (i = 0; i < NSIG; i++) {
- if (sigismember(&blocked_signals, i))
- signal(i, SIG_DFL);
- }
+/* Temporarily disabled due to 5423935 and 4946119. */
+#if 0
+ /* Ask the accountless process to exit. */
+ (void)job_assumes(j, runtime_kill(i_pid, SIGTERM) != -1);
+#endif
}
- sigprocmask(SIG_UNBLOCK, &blocked_signals, NULL);
-
- return r;
+out:
+ free(pids);
}
void
-machservice_resetport(struct jobcb *j, struct machservice *ms)
-{
- launchd_assumes(launchd_mport_close_recv(ms->port) == KERN_SUCCESS);
- launchd_assumes(launchd_mport_deallocate(ms->port) == KERN_SUCCESS);
- launchd_assumes(launchd_mport_create_recv(&ms->port) == KERN_SUCCESS);
- launchd_assumes(launchd_mport_make_send(ms->port) == KERN_SUCCESS);
-}
-
-struct machservice *
-machservice_new(struct jobcb *j, const char *name, mach_port_t *serviceport)
+job_postfork_test_user(job_t j)
{
- struct machservice *ms;
+ /* This function is all about 5201578 */
- if ((ms = calloc(1, sizeof(struct machservice) + strlen(name) + 1)) == NULL)
- return NULL;
+ const char *home_env_var = getenv("HOME");
+ const char *user_env_var = getenv("USER");
+ const char *logname_env_var = getenv("LOGNAME");
+ uid_t tmp_uid, local_uid = getuid();
+ gid_t tmp_gid, local_gid = getgid();
+ char shellpath[PATH_MAX];
+ char homedir[PATH_MAX];
+ char loginname[2000];
+ struct passwd *pwe;
- strcpy(ms->name, name);
- ms->job = j;
- if (*serviceport == MACH_PORT_NULL) {
- if (!launchd_assumes(launchd_mport_create_recv(&ms->port) == KERN_SUCCESS))
- goto out_bad;
+ if (!job_assumes(j, home_env_var && user_env_var && logname_env_var
+ && strcmp(user_env_var, logname_env_var) == 0)) {
+ goto out_bad;
+ }
- if (!launchd_assumes(launchd_mport_make_send(ms->port) == KERN_SUCCESS))
- goto out_bad2;
- *serviceport = ms->port;
- ms->isActive = false;
- ms->recv = true;
- } else {
- ms->port = *serviceport;
- ms->isActive = true;
+ if ((pwe = getpwnam(user_env_var)) == NULL) {
+ job_log(j, LOG_ERR, "The account \"%s\" has been deleted out from under us!", user_env_var);
+ goto out_bad;
}
- SLIST_INSERT_HEAD(&j->machservices, ms, sle);
+ /*
+ * We must copy the results of getpw*().
+ *
+ * Why? Because subsequent API calls may call getpw*() as a part of
+ * their implementation. Since getpw*() returns a [now thread scoped]
+ * global, we must therefore cache the results before continuing.
+ */
- job_log(j, LOG_INFO, "Mach service added: %s", name);
+ tmp_uid = pwe->pw_uid;
+ tmp_gid = pwe->pw_gid;
- return ms;
-out_bad2:
- launchd_assumes(launchd_mport_close_recv(ms->port) == KERN_SUCCESS);
-out_bad:
- free(ms);
- return NULL;
-}
+ strlcpy(shellpath, pwe->pw_shell, sizeof(shellpath));
+ strlcpy(loginname, pwe->pw_name, sizeof(loginname));
+ strlcpy(homedir, pwe->pw_dir, sizeof(homedir));
-bootstrap_status_t
-machservice_status(struct machservice *ms)
-{
- if (ms->isActive) {
- return BOOTSTRAP_STATUS_ACTIVE;
- } else if (ms->job->ondemand) {
- return BOOTSTRAP_STATUS_ON_DEMAND;
- } else {
- return BOOTSTRAP_STATUS_INACTIVE;
+ if (strcmp(loginname, logname_env_var) != 0) {
+ job_log(j, LOG_ERR, "The %s environmental variable changed out from under us!", "USER");
+ goto out_bad;
}
+ if (strcmp(homedir, home_env_var) != 0) {
+ job_log(j, LOG_ERR, "The %s environmental variable changed out from under us!", "HOME");
+ goto out_bad;
+ }
+ if (local_uid != tmp_uid) {
+ job_log(j, LOG_ERR, "The %cID of the account (%u) changed out from under us (%u)!",
+ 'U', tmp_uid, local_uid);
+ goto out_bad;
+ }
+ if (local_gid != tmp_gid) {
+ job_log(j, LOG_ERR, "The %cID of the account (%u) changed out from under us (%u)!",
+ 'G', tmp_gid, local_gid);
+ goto out_bad;
+ }
+
+ return;
+out_bad:
+#if 0
+ (void)job_assumes(j, runtime_kill(getppid(), SIGTERM) != -1);
+ _exit(EXIT_FAILURE);
+#else
+ job_log(j, LOG_WARNING, "In a future build of the OS, this error will be fatal.");
+#endif
}
void
-machservice_setup_options(launch_data_t obj, const char *key, void *context)
+job_postfork_become_user(job_t j)
{
- struct machservice *ms = context;
- mach_port_t mhp = mach_host_self();
- mach_port_t mts = mach_task_self();
- thread_state_flavor_t f = 0;
- int which_port;
- bool b;
-
-#if defined (__ppc__)
- f = PPC_THREAD_STATE64;
-#elif defined(__i386__)
- f = x86_THREAD_STATE;
-#endif
+ char loginname[2000];
+ char tmpdirpath[PATH_MAX];
+ char shellpath[PATH_MAX];
+ char homedir[PATH_MAX];
+ struct passwd *pwe;
+ size_t r;
+ gid_t desired_gid = -1;
+ uid_t desired_uid = -1;
+
+ if (getuid() != 0) {
+ return job_postfork_test_user(j);
+ }
- if (!launchd_assumes(mhp != MACH_PORT_NULL)) {
- return;
+ /*
+ * I contend that having UID == 0 and GID != 0 is of dubious value.
+ * Nevertheless, this used to work in Tiger. See: 5425348
+ */
+ if (j->groupname && !j->username) {
+ j->username = "root";
}
- switch (launch_data_get_type(obj)) {
+ if (j->username) {
+ if ((pwe = getpwnam(j->username)) == NULL) {
+ job_log(j, LOG_ERR, "getpwnam(\"%s\") failed", j->username);
+ _exit(EXIT_FAILURE);
+ }
+ } else if (j->mach_uid) {
+ if ((pwe = getpwuid(j->mach_uid)) == NULL) {
+ job_log(j, LOG_ERR, "getpwuid(\"%u\") failed", j->mach_uid);
+ job_log_pids_with_weird_uids(j);
+ _exit(EXIT_FAILURE);
+ }
+ } else {
+ return;
+ }
+
+ /*
+ * We must copy the results of getpw*().
+ *
+ * Why? Because subsequent API calls may call getpw*() as a part of
+ * their implementation. Since getpw*() returns a [now thread scoped]
+ * global, we must therefore cache the results before continuing.
+ */
+
+ desired_uid = pwe->pw_uid;
+ desired_gid = pwe->pw_gid;
+
+ strlcpy(shellpath, pwe->pw_shell, sizeof(shellpath));
+ strlcpy(loginname, pwe->pw_name, sizeof(loginname));
+ strlcpy(homedir, pwe->pw_dir, sizeof(homedir));
+
+ if (unlikely(pwe->pw_expire && time(NULL) >= pwe->pw_expire)) {
+ job_log(j, LOG_ERR, "Expired account");
+ _exit(EXIT_FAILURE);
+ }
+
+
+ if (unlikely(j->username && strcmp(j->username, loginname) != 0)) {
+ job_log(j, LOG_WARNING, "Suspicious setup: User \"%s\" maps to user: %s", j->username, loginname);
+ } else if (unlikely(j->mach_uid && (j->mach_uid != desired_uid))) {
+ job_log(j, LOG_WARNING, "Suspicious setup: UID %u maps to UID %u", j->mach_uid, desired_uid);
+ }
+
+ if (j->groupname) {
+ struct group *gre;
+
+ if (unlikely((gre = getgrnam(j->groupname)) == NULL)) {
+ job_log(j, LOG_ERR, "getgrnam(\"%s\") failed", j->groupname);
+ _exit(EXIT_FAILURE);
+ }
+
+ desired_gid = gre->gr_gid;
+ }
+
+ if (!job_assumes(j, setlogin(loginname) != -1)) {
+ _exit(EXIT_FAILURE);
+ }
+
+ if (!job_assumes(j, setgid(desired_gid) != -1)) {
+ _exit(EXIT_FAILURE);
+ }
+
+ /*
+ * The kernel team and the DirectoryServices team want initgroups()
+ * called after setgid(). See 4616864 for more information.
+ */
+
+ if (likely(!j->no_init_groups)) {
+ #if 1
+ if (!job_assumes(j, initgroups(loginname, desired_gid) != -1)) {
+ _exit(EXIT_FAILURE);
+ }
+ #else
+ /* Do our own little initgroups(). We do this to guarantee that we're
+ * always opted into dynamic group resolution in the kernel. initgroups(3)
+ * does not make this guarantee.
+ */
+ int groups[NGROUPS], ngroups;
+
+ /* A failure here isn't fatal, and we'll still get data we can use. */
+ (void)job_assumes(j, getgrouplist(j->username, desired_gid, groups, &ngroups) != -1);
+
+ if (!job_assumes(j, syscall(SYS_initgroups, ngroups, groups, desired_uid) != -1)) {
+ _exit(EXIT_FAILURE);
+ }
+ #endif
+ }
+
+ if (!job_assumes(j, setuid(desired_uid) != -1)) {
+ _exit(EXIT_FAILURE);
+ }
+
+ r = confstr(_CS_DARWIN_USER_TEMP_DIR, tmpdirpath, sizeof(tmpdirpath));
+
+ if (likely(r > 0 && r < sizeof(tmpdirpath))) {
+ setenv("TMPDIR", tmpdirpath, 0);
+ }
+
+ setenv("SHELL", shellpath, 0);
+ setenv("HOME", homedir, 0);
+ setenv("USER", loginname, 0);
+ setenv("LOGNAME", loginname, 0);
+}
+
+void
+job_setup_attributes(job_t j)
+{
+ struct limititem *li;
+ struct envitem *ei;
+
+ if (unlikely(j->setnice)) {
+ (void)job_assumes(j, setpriority(PRIO_PROCESS, 0, j->nice) != -1);
+ }
+
+ SLIST_FOREACH(li, &j->limits, sle) {
+ struct rlimit rl;
+
+ if (!job_assumes(j, getrlimit(li->which, &rl) != -1)) {
+ continue;
+ }
+
+ if (li->sethard) {
+ rl.rlim_max = li->lim.rlim_max;
+ }
+ if (li->setsoft) {
+ rl.rlim_cur = li->lim.rlim_cur;
+ }
+
+ if (setrlimit(li->which, &rl) == -1) {
+ job_log_error(j, LOG_WARNING, "setrlimit()");
+ }
+ }
+
+ if (unlikely(!j->inetcompat && j->session_create)) {
+ launchd_SessionCreate();
+ }
+
+ if (unlikely(j->low_pri_io)) {
+ (void)job_assumes(j, setiopolicy_np(IOPOL_TYPE_DISK, IOPOL_SCOPE_PROCESS, IOPOL_THROTTLE) != -1);
+ }
+ if (unlikely(j->rootdir)) {
+ (void)job_assumes(j, chroot(j->rootdir) != -1);
+ (void)job_assumes(j, chdir(".") != -1);
+ }
+
+ job_postfork_become_user(j);
+
+ if (unlikely(j->workingdir)) {
+ (void)job_assumes(j, chdir(j->workingdir) != -1);
+ }
+
+ if (unlikely(j->setmask)) {
+ umask(j->mask);
+ }
+
+ if (j->stdin_fd) {
+ (void)job_assumes(j, dup2(j->stdin_fd, STDIN_FILENO) != -1);
+ } else {
+ job_setup_fd(j, STDIN_FILENO, j->stdinpath, O_RDONLY|O_CREAT);
+ }
+ job_setup_fd(j, STDOUT_FILENO, j->stdoutpath, O_WRONLY|O_CREAT|O_APPEND);
+ job_setup_fd(j, STDERR_FILENO, j->stderrpath, O_WRONLY|O_CREAT|O_APPEND);
+
+ jobmgr_setup_env_from_other_jobs(j->mgr);
+
+ SLIST_FOREACH(ei, &j->env, sle) {
+ setenv(ei->key, ei->value, 1);
+ }
+
+ if (do_apple_internal_logging) {
+ setenv(LAUNCHD_DO_APPLE_INTERNAL_LOGGING, "true", 1);
+ }
+
+#if !TARGET_OS_EMBEDDED
+ if (j->jetsam_properties) {
+ (void)job_assumes(j, proc_setpcontrol(PROC_SETPC_TERMINATE) == 0);
+ }
+#endif
+
+#if TARGET_OS_EMBEDDED
+ if (j->main_thread_priority != 0) {
+ struct sched_param params;
+ bzero(¶ms, sizeof(params));
+ params.sched_priority = j->main_thread_priority;
+ (void)job_assumes(j, pthread_setschedparam(pthread_self(), SCHED_OTHER, ¶ms) != -1);
+ }
+#endif
+
+ /*
+ * We'd like to call setsid() unconditionally, but we have reason to
+ * believe that prevents launchd from being able to send signals to
+ * setuid children. We'll settle for process-groups.
+ */
+ if (getppid() != 1) {
+ (void)job_assumes(j, setpgid(0, 0) != -1);
+ } else {
+ (void)job_assumes(j, setsid() != -1);
+ }
+}
+
+void
+job_setup_fd(job_t j, int target_fd, const char *path, int flags)
+{
+ int fd;
+
+ if (!path) {
+ return;
+ }
+
+ if ((fd = open(path, flags|O_NOCTTY, DEFFILEMODE)) == -1) {
+ job_log_error(j, LOG_WARNING, "open(\"%s\", ...)", path);
+ return;
+ }
+
+ (void)job_assumes(j, dup2(fd, target_fd) != -1);
+ (void)job_assumes(j, runtime_close(fd) == 0);
+}
+
+int
+dir_has_files(job_t j, const char *path)
+{
+ DIR *dd = opendir(path);
+ struct dirent *de;
+ bool r = 0;
+
+ if (unlikely(!dd)) {
+ return -1;
+ }
+
+ while ((de = readdir(dd))) {
+ if (strcmp(de->d_name, ".") && strcmp(de->d_name, "..")) {
+ r = 1;
+ break;
+ }
+ }
+
+ (void)job_assumes(j, closedir(dd) == 0);
+ return r;
+}
+
+void
+calendarinterval_setalarm(job_t j, struct calendarinterval *ci)
+{
+ struct calendarinterval *ci_iter, *ci_prev = NULL;
+ time_t later, head_later;
+
+ later = cronemu(ci->when.tm_mon, ci->when.tm_mday, ci->when.tm_hour, ci->when.tm_min);
+
+ if (ci->when.tm_wday != -1) {
+ time_t otherlater = cronemu_wday(ci->when.tm_wday, ci->when.tm_hour, ci->when.tm_min);
+
+ if (ci->when.tm_mday == -1) {
+ later = otherlater;
+ } else {
+ later = later < otherlater ? later : otherlater;
+ }
+ }
+
+ ci->when_next = later;
+
+ LIST_FOREACH(ci_iter, &sorted_calendar_events, global_sle) {
+ if (ci->when_next < ci_iter->when_next) {
+ LIST_INSERT_BEFORE(ci_iter, ci, global_sle);
+ break;
+ }
+
+ ci_prev = ci_iter;
+ }
+
+ if (ci_iter == NULL) {
+ /* ci must want to fire after every other timer, or there are no timers */
+
+ if (LIST_EMPTY(&sorted_calendar_events)) {
+ LIST_INSERT_HEAD(&sorted_calendar_events, ci, global_sle);
+ } else {
+ LIST_INSERT_AFTER(ci_prev, ci, global_sle);
+ }
+ }
+
+ head_later = LIST_FIRST(&sorted_calendar_events)->when_next;
+
+ if (job_assumes(j, kevent_mod((uintptr_t)&sorted_calendar_events, EVFILT_TIMER, EV_ADD, NOTE_ABSOLUTE|NOTE_SECONDS, head_later, root_jobmgr) != -1)) {
+ char time_string[100];
+ size_t time_string_len;
+
+ ctime_r(&later, time_string);
+ time_string_len = strlen(time_string);
+
+ if (likely(time_string_len && time_string[time_string_len - 1] == '\n')) {
+ time_string[time_string_len - 1] = '\0';
+ }
+
+ job_log(j, LOG_INFO, "Scheduled to run again at %s", time_string);
+ }
+}
+
+void
+extract_rcsid_substr(const char *i, char *o, size_t osz)
+{
+ char *rcs_rev_tmp = strchr(i, ' ');
+
+ if (!rcs_rev_tmp) {
+ strlcpy(o, i, osz);
+ } else {
+ strlcpy(o, rcs_rev_tmp + 1, osz);
+ rcs_rev_tmp = strchr(o, ' ');
+ if (rcs_rev_tmp) {
+ *rcs_rev_tmp = '\0';
+ }
+ }
+}
+
+void
+jobmgr_log_bug(jobmgr_t jm, unsigned int line)
+{
+ static const char *file;
+ int saved_errno = errno;
+ char buf[100];
+
+ runtime_ktrace1(RTKT_LAUNCHD_BUG);
+
+ extract_rcsid_substr(__rcs_file_version__, buf, sizeof(buf));
+
+ if (!file) {
+ file = strrchr(__FILE__, '/');
+ if (!file) {
+ file = __FILE__;
+ } else {
+ file += 1;
+ }
+ }
+
+ /* the only time 'jm' should not be set is if setting up the first bootstrap fails for some reason */
+ if (likely(jm)) {
+ jobmgr_log(jm, LOG_NOTICE, "Bug: %s:%u (%s):%u", file, line, buf, saved_errno);
+ } else {
+ runtime_syslog(LOG_NOTICE, "Bug: %s:%u (%s):%u", file, line, buf, saved_errno);
+ }
+}
+
+void
+job_log_bug(job_t j, unsigned int line)
+{
+ static const char *file;
+ int saved_errno = errno;
+ char buf[100];
+
+ runtime_ktrace1(RTKT_LAUNCHD_BUG);
+
+ extract_rcsid_substr(__rcs_file_version__, buf, sizeof(buf));
+
+ if (!file) {
+ file = strrchr(__FILE__, '/');
+ if (!file) {
+ file = __FILE__;
+ } else {
+ file += 1;
+ }
+ }
+
+ if (likely(j)) {
+ job_log(j, LOG_NOTICE, "Bug: %s:%u (%s):%u", file, line, buf, saved_errno);
+ } else {
+ runtime_syslog(LOG_NOTICE, "Bug: %s:%u (%s):%u", file, line, buf, saved_errno);
+ }
+}
+
+void
+job_logv(job_t j, int pri, int err, const char *msg, va_list ap)
+{
+ const char *label2use = j ? j->label : "com.apple.launchd.NULL";
+ const char *mgr2use = j ? j->mgr->name : "NULL";
+ struct runtime_syslog_attr attr = { g_my_label, label2use, mgr2use, pri, getuid(), getpid(), j ? j->p : 0 };
+ char *newmsg;
+ int oldmask = 0;
+ size_t newmsgsz;
+
+ /*
+ * Hack: If bootstrap_port is set, we must be on the child side of a
+ * fork(), but before the exec*(). Let's route the log message back to
+ * launchd proper.
+ */
+ if (bootstrap_port) {
+ return _vproc_logv(pri, err, msg, ap);
+ }
+
+ newmsgsz = strlen(msg) + 200;
+ newmsg = alloca(newmsgsz);
+
+ if (err) {
+#if !TARGET_OS_EMBEDDED
+ snprintf(newmsg, newmsgsz, "%s: %s", msg, strerror(err));
+#else
+ snprintf(newmsg, newmsgsz, "(%s) %s: %s", label2use, msg, strerror(err));
+#endif
+ } else {
+#if !TARGET_OS_EMBEDDED
+ snprintf(newmsg, newmsgsz, "%s", msg);
+#else
+ snprintf(newmsg, newmsgsz, "(%s) %s", label2use, msg);
+#endif
+ }
+
+ if (j && unlikely(j->debug)) {
+ oldmask = setlogmask(LOG_UPTO(LOG_DEBUG));
+ }
+
+ runtime_vsyslog(&attr, newmsg, ap);
+
+ if (j && unlikely(j->debug)) {
+ setlogmask(oldmask);
+ }
+}
+
+void
+job_log_error(job_t j, int pri, const char *msg, ...)
+{
+ va_list ap;
+
+ va_start(ap, msg);
+ job_logv(j, pri, errno, msg, ap);
+ va_end(ap);
+}
+
+void
+job_log(job_t j, int pri, const char *msg, ...)
+{
+ va_list ap;
+
+ va_start(ap, msg);
+ job_logv(j, pri, 0, msg, ap);
+ va_end(ap);
+}
+
+#if 0
+void
+jobmgr_log_error(jobmgr_t jm, int pri, const char *msg, ...)
+{
+ va_list ap;
+
+ va_start(ap, msg);
+ jobmgr_logv(jm, pri, errno, msg, ap);
+ va_end(ap);
+}
+#endif
+
+void
+jobmgr_log(jobmgr_t jm, int pri, const char *msg, ...)
+{
+ va_list ap;
+
+ va_start(ap, msg);
+ jobmgr_logv(jm, pri, 0, msg, ap);
+ va_end(ap);
+}
+
+void
+jobmgr_logv(jobmgr_t jm, int pri, int err, const char *msg, va_list ap)
+{
+ char *newmsg;
+ char *newname;
+ size_t i, o, jmname_len = strlen(jm->name), newmsgsz;
+
+ newname = alloca((jmname_len + 1) * 2);
+ newmsgsz = (jmname_len + 1) * 2 + strlen(msg) + 100;
+ newmsg = alloca(newmsgsz);
+
+ for (i = 0, o = 0; i < jmname_len; i++, o++) {
+ if (jm->name[i] == '%') {
+ newname[o] = '%';
+ o++;
+ }
+ newname[o] = jm->name[i];
+ }
+ newname[o] = '\0';
+
+ if (err) {
+ snprintf(newmsg, newmsgsz, "%s: %s: %s", newname, msg, strerror(err));
+ } else {
+ snprintf(newmsg, newmsgsz, "%s: %s", newname, msg);
+ }
+
+ if (jm->parentmgr) {
+ jobmgr_logv(jm->parentmgr, pri, 0, newmsg, ap);
+ } else {
+ struct runtime_syslog_attr attr = { g_my_label, g_my_label, jm->name, pri, getuid(), getpid(), getpid() };
+
+ runtime_vsyslog(&attr, newmsg, ap);
+ }
+}
+
+void
+semaphoreitem_ignore(job_t j, struct semaphoreitem *si)
+{
+ if (si->fd != -1) {
+ job_log(j, LOG_DEBUG, "Ignoring Vnode: %d", si->fd);
+ (void)job_assumes(j, kevent_mod(si->fd, EVFILT_VNODE, EV_DELETE, 0, 0, NULL) != -1);
+ }
+}
+
+void
+semaphoreitem_watch(job_t j, struct semaphoreitem *si)
+{
+ char *parentdir, tmp_path[PATH_MAX];
+ int saved_errno = 0;
+ int fflags = NOTE_DELETE|NOTE_RENAME;
+
+ switch (si->why) {
+ case DIR_NOT_EMPTY:
+ case PATH_CHANGES:
+ fflags |= NOTE_ATTRIB|NOTE_LINK;
+ /* fall through */
+ case PATH_EXISTS:
+ fflags |= NOTE_REVOKE|NOTE_EXTEND|NOTE_WRITE;
+ /* fall through */
+ case PATH_MISSING:
+ break;
+ default:
+ return;
+ }
+
+ /* dirname() may modify tmp_path */
+ strlcpy(tmp_path, si->what, sizeof(tmp_path));
+
+ if (!job_assumes(j, (parentdir = dirname(tmp_path)))) {
+ return;
+ }
+
+ /* See 5321044 for why we do the do-while loop and 5415523 for why ENOENT is checked */
+ do {
+ if (si->fd == -1) {
+ struct stat sb;
+ if (stat(si->what, &sb) == 0) {
+ /* If we're watching a character or block device, only watch the parent directory.
+ * See rdar://problem/6489900 for the gory details. Basically, holding an open file
+ * descriptor to a devnode could end up (a) blocking us on open(2) until someone else
+ * open(2)s the file (like a character device that waits for a carrier signal) or
+ * (b) preventing other processes from obtaining an exclusive lock on the file, even
+ * though we're opening it with O_EVTONLY.
+ *
+ * The main point of contention is that O_EVTONLY doesn't actually mean "event only".
+ * It means "Don't prevent unmounts of this descriptor's volume". We work around this
+ * for dev nodes by only watching the parent directory and stat(2)ing our desired file
+ * each time the parent changes to see if it appeared or disappeared.
+ */
+ if (S_ISREG(sb.st_mode) || S_ISDIR(sb.st_mode)) {
+ si->fd = _fd(open(si->what, O_EVTONLY | O_NOCTTY | O_NONBLOCK));
+ }
+ }
+
+ if (si->fd == -1) {
+ si->watching_parent = job_assumes(j, (si->fd = _fd(open(parentdir, O_EVTONLY | O_NOCTTY | O_NONBLOCK))) != -1);
+ } else {
+ si->watching_parent = false;
+ }
+ }
+
+ if (si->fd == -1) {
+ return job_log_error(j, LOG_ERR, "Path monitoring failed on \"%s\"", si->what);
+ }
+
+ job_log(j, LOG_DEBUG, "Watching %svnode (%s): %d", si->watching_parent ? "parent ": "", si->what, si->fd);
+
+ if (kevent_mod(si->fd, EVFILT_VNODE, EV_ADD, fflags, 0, j) == -1) {
+ saved_errno = errno;
+ /*
+ * The FD can be revoked between the open() and kevent().
+ * This is similar to the inability for kevents to be
+ * attached to short lived zombie processes after fork()
+ * but before kevent().
+ */
+ (void)job_assumes(j, runtime_close(si->fd) == 0);
+ si->fd = -1;
+ }
+ } while (unlikely((si->fd == -1) && (saved_errno == ENOENT)));
+
+ if (saved_errno == ENOTSUP) {
+ /*
+ * 3524219 NFS needs kqueue support
+ * 4124079 VFS needs generic kqueue support
+ * 5226811 EVFILT: Launchd EVFILT_VNODE doesn't work on /dev
+ */
+ job_log(j, LOG_DEBUG, "Falling back to polling for path: %s", si->what);
+
+ if (!j->poll_for_vfs_changes) {
+ j->poll_for_vfs_changes = true;
+ (void)job_assumes(j, kevent_mod((uintptr_t)&j->semaphores, EVFILT_TIMER, EV_ADD, NOTE_SECONDS, 3, j) != -1);
+ }
+ }
+}
+
+void
+semaphoreitem_callback(job_t j, struct kevent *kev)
+{
+ char invalidation_reason[100] = "";
+ struct semaphoreitem *si;
+
+ SLIST_FOREACH(si, &j->semaphores, sle) {
+ switch (si->why) {
+ case PATH_CHANGES:
+ case PATH_EXISTS:
+ case PATH_MISSING:
+ case DIR_NOT_EMPTY:
+ job_log(j, LOG_DEBUG, "P%s changed (%u): %s", si->watching_parent ? "arent path" : "ath", si->why, si->what);
+ break;
+ default:
+ continue;
+ }
+
+ if (si->fd == (int)kev->ident) {
+ break;
+ }
+ }
+
+ if (!job_assumes(j, si != NULL)) {
+ return;
+ }
+
+ if (NOTE_DELETE & kev->fflags) {
+ strcat(invalidation_reason, "deleted");
+ }
+
+ if (NOTE_RENAME & kev->fflags) {
+ if (invalidation_reason[0]) {
+ strcat(invalidation_reason, "/renamed");
+ } else {
+ strcat(invalidation_reason, "renamed");
+ }
+ }
+
+ if (NOTE_REVOKE & kev->fflags) {
+ if (invalidation_reason[0]) {
+ strcat(invalidation_reason, "/revoked");
+ } else {
+ strcat(invalidation_reason, "revoked");
+ }
+ }
+
+ if (invalidation_reason[0]) {
+ job_log(j, LOG_DEBUG, "Path %s: %s", invalidation_reason, si->what);
+ (void)job_assumes(j, runtime_close(si->fd) == 0);
+ si->fd = -1; /* this will get fixed in semaphoreitem_watch() */
+ }
+
+ if (!si->watching_parent) {
+ if (si->why == PATH_CHANGES) {
+ j->start_pending = true;
+ } else {
+ semaphoreitem_watch(j, si);
+ }
+ } else { /* Something happened to the parent directory. See if our target file appeared. */
+ if (!invalidation_reason[0]) {
+ (void)job_assumes(j, runtime_close(si->fd) == 0);
+ si->fd = -1; /* this will get fixed in semaphoreitem_watch() */
+ semaphoreitem_watch(j, si);
+ }
+ /* Need to think about what should happen if the parent directory goes invalid. */
+ }
+
+ job_dispatch(j, false);
+}
+
+struct cal_dict_walk {
+ job_t j;
+ struct tm tmptm;
+};
+
+void
+calendarinterval_new_from_obj_dict_walk(launch_data_t obj, const char *key, void *context)
+{
+ struct cal_dict_walk *cdw = context;
+ struct tm *tmptm = &cdw->tmptm;
+ job_t j = cdw->j;
+ int64_t val;
+
+ if (unlikely(LAUNCH_DATA_INTEGER != launch_data_get_type(obj))) {
+ /* hack to let caller know something went wrong */
+ tmptm->tm_sec = -1;
+ return;
+ }
+
+ val = launch_data_get_integer(obj);
+
+ if (val < 0) {
+ job_log(j, LOG_WARNING, "The interval for key \"%s\" is less than zero.", key);
+ } else if (strcasecmp(key, LAUNCH_JOBKEY_CAL_MINUTE) == 0) {
+ if (val > 59) {
+ job_log(j, LOG_WARNING, "The interval for key \"%s\" is not between 0 and 59 (inclusive).", key);
+ tmptm->tm_sec = -1;
+ } else {
+ tmptm->tm_min = (typeof(tmptm->tm_min)) val;
+ }
+ } else if (strcasecmp(key, LAUNCH_JOBKEY_CAL_HOUR) == 0) {
+ if (val > 23) {
+ job_log(j, LOG_WARNING, "The interval for key \"%s\" is not between 0 and 23 (inclusive).", key);
+ tmptm->tm_sec = -1;
+ } else {
+ tmptm->tm_hour = (typeof(tmptm->tm_hour)) val;
+ }
+ } else if (strcasecmp(key, LAUNCH_JOBKEY_CAL_DAY) == 0) {
+ if (val < 1 || val > 31) {
+ job_log(j, LOG_WARNING, "The interval for key \"%s\" is not between 1 and 31 (inclusive).", key);
+ tmptm->tm_sec = -1;
+ } else {
+ tmptm->tm_mday = (typeof(tmptm->tm_mday)) val;
+ }
+ } else if (strcasecmp(key, LAUNCH_JOBKEY_CAL_WEEKDAY) == 0) {
+ if (val > 7) {
+ job_log(j, LOG_WARNING, "The interval for key \"%s\" is not between 0 and 7 (inclusive).", key);
+ tmptm->tm_sec = -1;
+ } else {
+ tmptm->tm_wday = (typeof(tmptm->tm_wday)) val;
+ }
+ } else if (strcasecmp(key, LAUNCH_JOBKEY_CAL_MONTH) == 0) {
+ if (val > 12) {
+ job_log(j, LOG_WARNING, "The interval for key \"%s\" is not between 0 and 12 (inclusive).", key);
+ tmptm->tm_sec = -1;
+ } else {
+ tmptm->tm_mon = (typeof(tmptm->tm_mon)) val;
+ tmptm->tm_mon -= 1; /* 4798263 cron compatibility */
+ }
+ }
+}
+
+bool
+calendarinterval_new_from_obj(job_t j, launch_data_t obj)
+{
+ struct cal_dict_walk cdw;
+
+ cdw.j = j;
+ memset(&cdw.tmptm, 0, sizeof(0));
+
+ cdw.tmptm.tm_min = -1;
+ cdw.tmptm.tm_hour = -1;
+ cdw.tmptm.tm_mday = -1;
+ cdw.tmptm.tm_wday = -1;
+ cdw.tmptm.tm_mon = -1;
+
+ if (!job_assumes(j, obj != NULL)) {
+ return false;
+ }
+
+ if (unlikely(LAUNCH_DATA_DICTIONARY != launch_data_get_type(obj))) {
+ return false;
+ }
+
+ launch_data_dict_iterate(obj, calendarinterval_new_from_obj_dict_walk, &cdw);
+
+ if (unlikely(cdw.tmptm.tm_sec == -1)) {
+ return false;
+ }
+
+ return calendarinterval_new(j, &cdw.tmptm);
+}
+
+bool
+calendarinterval_new(job_t j, struct tm *w)
+{
+ struct calendarinterval *ci = calloc(1, sizeof(struct calendarinterval));
+
+ if (!job_assumes(j, ci != NULL)) {
+ return false;
+ }
+
+ ci->when = *w;
+ ci->job = j;
+
+ SLIST_INSERT_HEAD(&j->cal_intervals, ci, sle);
+
+ calendarinterval_setalarm(j, ci);
+
+ runtime_add_weak_ref();
+
+ return true;
+}
+
+void
+calendarinterval_delete(job_t j, struct calendarinterval *ci)
+{
+ SLIST_REMOVE(&j->cal_intervals, ci, calendarinterval, sle);
+ LIST_REMOVE(ci, global_sle);
+
+ free(ci);
+
+ runtime_del_weak_ref();
+}
+
+void
+calendarinterval_sanity_check(void)
+{
+ struct calendarinterval *ci = LIST_FIRST(&sorted_calendar_events);
+ time_t now = time(NULL);
+
+ if (unlikely(ci && (ci->when_next < now))) {
+ (void)jobmgr_assumes(root_jobmgr, raise(SIGUSR1) != -1);
+ }
+}
+
+void
+calendarinterval_callback(void)
+{
+ struct calendarinterval *ci, *ci_next;
+ time_t now = time(NULL);
+
+ LIST_FOREACH_SAFE(ci, &sorted_calendar_events, global_sle, ci_next) {
+ job_t j = ci->job;
+
+ if (ci->when_next > now) {
+ break;
+ }
+
+ LIST_REMOVE(ci, global_sle);
+ calendarinterval_setalarm(j, ci);
+
+ j->start_pending = true;
+ job_dispatch(j, false);
+ }
+}
+
+bool
+socketgroup_new(job_t j, const char *name, int *fds, size_t fd_cnt, bool junkfds)
+{
+ struct socketgroup *sg = calloc(1, sizeof(struct socketgroup) + strlen(name) + 1);
+
+ if (!job_assumes(j, sg != NULL)) {
+ return false;
+ }
+
+ sg->fds = calloc(1, fd_cnt * sizeof(int));
+ sg->fd_cnt = fd_cnt;
+ sg->junkfds = junkfds;
+
+ if (!job_assumes(j, sg->fds != NULL)) {
+ free(sg);
+ return false;
+ }
+
+ memcpy(sg->fds, fds, fd_cnt * sizeof(int));
+ strcpy(sg->name_init, name);
+
+ SLIST_INSERT_HEAD(&j->sockets, sg, sle);
+
+ runtime_add_weak_ref();
+
+ return true;
+}
+
+void
+socketgroup_delete(job_t j, struct socketgroup *sg)
+{
+ unsigned int i;
+
+ for (i = 0; i < sg->fd_cnt; i++) {
+#if 0
+ struct sockaddr_storage ss;
+ struct sockaddr_un *sun = (struct sockaddr_un *)&ss;
+ socklen_t ss_len = sizeof(ss);
+
+ /* 5480306 */
+ if (job_assumes(j, getsockname(sg->fds[i], (struct sockaddr *)&ss, &ss_len) != -1)
+ && job_assumes(j, ss_len > 0) && (ss.ss_family == AF_UNIX)) {
+ (void)job_assumes(j, unlink(sun->sun_path) != -1);
+ /* We might conditionally need to delete a directory here */
+ }
+#endif
+ (void)job_assumes(j, runtime_close(sg->fds[i]) != -1);
+ }
+
+ SLIST_REMOVE(&j->sockets, sg, socketgroup, sle);
+
+ free(sg->fds);
+ free(sg);
+
+ runtime_del_weak_ref();
+}
+
+void
+socketgroup_kevent_mod(job_t j, struct socketgroup *sg, bool do_add)
+{
+ struct kevent kev[sg->fd_cnt];
+ char buf[10000];
+ unsigned int i, buf_off = 0;
+
+ if (unlikely(sg->junkfds)) {
+ return;
+ }
+
+ for (i = 0; i < sg->fd_cnt; i++) {
+ EV_SET(&kev[i], sg->fds[i], EVFILT_READ, do_add ? EV_ADD : EV_DELETE, 0, 0, j);
+ buf_off += snprintf(buf + buf_off, sizeof(buf) - buf_off, " %d", sg->fds[i]);
+ }
+
+ job_log(j, LOG_DEBUG, "%s Sockets:%s", do_add ? "Watching" : "Ignoring", buf);
+
+ (void)job_assumes(j, kevent_bulk_mod(kev, sg->fd_cnt) != -1);
+
+ for (i = 0; i < sg->fd_cnt; i++) {
+ (void)job_assumes(j, kev[i].flags & EV_ERROR);
+ errno = (typeof(errno)) kev[i].data;
+ (void)job_assumes(j, kev[i].data == 0);
+ }
+}
+
+void
+socketgroup_ignore(job_t j, struct socketgroup *sg)
+{
+ socketgroup_kevent_mod(j, sg, false);
+}
+
+void
+socketgroup_watch(job_t j, struct socketgroup *sg)
+{
+ socketgroup_kevent_mod(j, sg, true);
+}
+
+void
+socketgroup_callback(job_t j)
+{
+ job_dispatch(j, true);
+}
+
+bool
+envitem_new(job_t j, const char *k, const char *v, bool global, bool one_shot)
+{
+ struct envitem *ei = calloc(1, sizeof(struct envitem) + strlen(k) + 1 + strlen(v) + 1);
+
+ if (!job_assumes(j, ei != NULL)) {
+ return false;
+ }
+
+ strcpy(ei->key_init, k);
+ ei->value = ei->key_init + strlen(k) + 1;
+ strcpy(ei->value, v);
+ ei->one_shot = one_shot;
+
+ if (global) {
+ if (SLIST_EMPTY(&j->global_env)) {
+ LIST_INSERT_HEAD(&j->mgr->global_env_jobs, j, global_env_sle);
+ }
+ SLIST_INSERT_HEAD(&j->global_env, ei, sle);
+ } else {
+ SLIST_INSERT_HEAD(&j->env, ei, sle);
+ }
+
+ job_log(j, LOG_DEBUG, "Added environmental variable: %s=%s", k, v);
+
+ return true;
+}
+
+void
+envitem_delete(job_t j, struct envitem *ei, bool global)
+{
+ if (global) {
+ SLIST_REMOVE(&j->global_env, ei, envitem, sle);
+ if (SLIST_EMPTY(&j->global_env)) {
+ LIST_REMOVE(j, global_env_sle);
+ }
+ } else {
+ SLIST_REMOVE(&j->env, ei, envitem, sle);
+ }
+
+ free(ei);
+}
+
+void
+envitem_setup(launch_data_t obj, const char *key, void *context)
+{
+ job_t j = context;
+
+ if (launch_data_get_type(obj) != LAUNCH_DATA_STRING) {
+ return;
+ }
+
+ if (strncmp(LAUNCHD_TRUSTED_FD_ENV, key, sizeof(LAUNCHD_TRUSTED_FD_ENV) - 1) != 0) {
+ envitem_new(j, key, launch_data_get_string(obj), j->importing_global_env, false);
+ } else {
+ job_log(j, LOG_DEBUG, "Ignoring reserved environmental variable: %s", key);
+ }
+}
+
+void
+envitem_setup_one_shot(launch_data_t obj, const char *key, void *context)
+{
+ job_t j = context;
+
+ if (launch_data_get_type(obj) != LAUNCH_DATA_STRING) {
+ return;
+ }
+
+ if (strncmp(LAUNCHD_TRUSTED_FD_ENV, key, sizeof(LAUNCHD_TRUSTED_FD_ENV) - 1) != 0) {
+ envitem_new(j, key, launch_data_get_string(obj), j->importing_global_env, true);
+ } else {
+ job_log(j, LOG_DEBUG, "Ignoring reserved environmental variable: %s", key);
+ }
+}
+
+bool
+limititem_update(job_t j, int w, rlim_t r)
+{
+ struct limititem *li;
+
+ SLIST_FOREACH(li, &j->limits, sle) {
+ if (li->which == w) {
+ break;
+ }
+ }
+
+ if (li == NULL) {
+ li = calloc(1, sizeof(struct limititem));
+
+ if (!job_assumes(j, li != NULL)) {
+ return false;
+ }
+
+ SLIST_INSERT_HEAD(&j->limits, li, sle);
+
+ li->which = w;
+ }
+
+ if (j->importing_hard_limits) {
+ li->lim.rlim_max = r;
+ li->sethard = true;
+ } else {
+ li->lim.rlim_cur = r;
+ li->setsoft = true;
+ }
+
+ return true;
+}
+
+void
+limititem_delete(job_t j, struct limititem *li)
+{
+ SLIST_REMOVE(&j->limits, li, limititem, sle);
+
+ free(li);
+}
+
+#if HAVE_SANDBOX
+void
+seatbelt_setup_flags(launch_data_t obj, const char *key, void *context)
+{
+ job_t j = context;
+
+ if (launch_data_get_type(obj) != LAUNCH_DATA_BOOL) {
+ job_log(j, LOG_WARNING, "Sandbox flag value must be boolean: %s", key);
+ return;
+ }
+
+ if (launch_data_get_bool(obj) == false) {
+ return;
+ }
+
+ if (strcasecmp(key, LAUNCH_JOBKEY_SANDBOX_NAMED) == 0) {
+ j->seatbelt_flags |= SANDBOX_NAMED;
+ }
+}
+#endif
+
+void
+limititem_setup(launch_data_t obj, const char *key, void *context)
+{
+ job_t j = context;
+ size_t i, limits_cnt = (sizeof(launchd_keys2limits) / sizeof(launchd_keys2limits[0]));
+ rlim_t rl;
+
+ if (launch_data_get_type(obj) != LAUNCH_DATA_INTEGER) {
+ return;
+ }
+
+ rl = launch_data_get_integer(obj);
+
+ for (i = 0; i < limits_cnt; i++) {
+ if (strcasecmp(launchd_keys2limits[i].key, key) == 0) {
+ break;
+ }
+ }
+
+ if (i == limits_cnt) {
+ return;
+ }
+
+ limititem_update(j, launchd_keys2limits[i].val, rl);
+}
+
+bool
+job_useless(job_t j)
+{
+ if ((j->legacy_LS_job || j->only_once) && j->start_time != 0) {
+ if (j->legacy_LS_job && j->j_port) {
+ return false;
+ }
+ job_log(j, LOG_INFO, "Exited. Was only configured to run once.");
+ return true;
+ } else if (j->removal_pending) {
+ job_log(j, LOG_DEBUG, "Exited while removal was pending.");
+ return true;
+ } else if (j->shutdown_monitor) {
+ return false;
+ } else if (j->mgr->shutting_down) {
+ job_log(j, LOG_DEBUG, "Exited while shutdown in progress. Processes remaining: %lu/%lu", total_children, total_anon_children);
+ if (total_children == 0 && !j->anonymous) {
+ job_log(j, LOG_DEBUG | LOG_CONSOLE, "Job was last to exit during shutdown of: %s.", j->mgr->name);
+ }
+ return true;
+ } else if (j->legacy_mach_job) {
+ if (SLIST_EMPTY(&j->machservices)) {
+ job_log(j, LOG_INFO, "Garbage collecting");
+ return true;
+ } else if (!j->checkedin) {
+ job_log(j, LOG_WARNING, "Failed to check-in!");
+ return true;
+ }
+ } else {
+ /* If the job's executable does not have any valid architectures (for
+ * example, if it's a PowerPC-only job), then we don't even bother
+ * trying to relaunch it, as we have no reasonable expectation that
+ * the situation will change.
+ *
+ * <rdar://problem/9106979>
+ */
+ if (!j->did_exec && WEXITSTATUS(j->last_exit_status) == EBADARCH) {
+ job_log(j, LOG_ERR, "Job executable does not contain supported architectures. Unloading it. Its plist should be removed.");
+ return true;
+ }
+ }
+
+ return false;
+}
+
+bool
+job_keepalive(job_t j)
+{
+ mach_msg_type_number_t statusCnt;
+ mach_port_status_t status;
+ struct semaphoreitem *si;
+ struct machservice *ms;
+ struct stat sb;
+ bool good_exit = (WIFEXITED(j->last_exit_status) && WEXITSTATUS(j->last_exit_status) == 0);
+ bool is_not_kextd = (do_apple_internal_logging || (strcmp(j->label, "com.apple.kextd") != 0));
+
+ if (unlikely(j->mgr->shutting_down)) {
+ return false;
+ }
+
+ /*
+ * 5066316
+ *
+ * We definitely need to revisit this after Leopard ships. Please see
+ * launchctl.c for the other half of this hack.
+ */
+ if (unlikely((j->mgr->global_on_demand_cnt > 0) && is_not_kextd)) {
+ return false;
+ }
+
+ if (unlikely(j->needs_kickoff)) {
+ job_log(j, LOG_DEBUG, "KeepAlive check: Job needs to be kicked off on-demand before KeepAlive sets in.");
+ return false;
+ }
+
+ if (j->start_pending) {
+ job_log(j, LOG_DEBUG, "KeepAlive check: Pent-up non-IPC launch criteria.");
+ return true;
+ }
+
+ if (!j->ondemand) {
+ job_log(j, LOG_DEBUG, "KeepAlive check: job configured to run continuously.");
+ return true;
+ }
+
+ SLIST_FOREACH(ms, &j->machservices, sle) {
+ statusCnt = MACH_PORT_RECEIVE_STATUS_COUNT;
+ if (mach_port_get_attributes(mach_task_self(), ms->port, MACH_PORT_RECEIVE_STATUS,
+ (mach_port_info_t)&status, &statusCnt) != KERN_SUCCESS) {
+ continue;
+ }
+ if (status.mps_msgcount) {
+ job_log(j, LOG_DEBUG, "KeepAlive check: %d queued Mach messages on service: %s",
+ status.mps_msgcount, ms->name);
+ return true;
+ }
+ }
+
+ /* TODO: Coalesce external events and semaphore items, since they're basically
+ * the same thing.
+ */
+ struct externalevent *ei = NULL;
+ LIST_FOREACH(ei, &j->events, job_le) {
+ if (ei->state == ei->wanted_state) {
+ return true;
+ }
+ }
+
+ SLIST_FOREACH(si, &j->semaphores, sle) {
+ bool wanted_state = false;
+ int qdir_file_cnt;
+ job_t other_j;
+
+ switch (si->why) {
+ case NETWORK_UP:
+ wanted_state = true;
+ case NETWORK_DOWN:
+ if (network_up == wanted_state) {
+ job_log(j, LOG_DEBUG, "KeepAlive: The network is %s.", wanted_state ? "up" : "down");
+ return true;
+ }
+ break;
+ case SUCCESSFUL_EXIT:
+ wanted_state = true;
+ case FAILED_EXIT:
+ if (good_exit == wanted_state) {
+ job_log(j, LOG_DEBUG, "KeepAlive: The exit state was %s.", wanted_state ? "successful" : "failure");
+ return true;
+ }
+ break;
+ case CRASHED:
+ wanted_state = true;
+ case DID_NOT_CRASH:
+ if (j->crashed == wanted_state) {
+ return true;
+ }
+ break;
+ case OTHER_JOB_ENABLED:
+ wanted_state = true;
+ case OTHER_JOB_DISABLED:
+ if ((bool)job_find(NULL, si->what) == wanted_state) {
+ job_log(j, LOG_DEBUG, "KeepAlive: The following job is %s: %s", wanted_state ? "enabled" : "disabled", si->what);
+ return true;
+ }
+ break;
+ case OTHER_JOB_ACTIVE:
+ wanted_state = true;
+ case OTHER_JOB_INACTIVE:
+ if ((other_j = job_find(NULL, si->what))) {
+ if ((bool)other_j->p == wanted_state) {
+ job_log(j, LOG_DEBUG, "KeepAlive: The following job is %s: %s", wanted_state ? "active" : "inactive", si->what);
+ return true;
+ }
+ }
+ break;
+ case PATH_EXISTS:
+ wanted_state = true;
+ case PATH_MISSING:
+ if ((bool)(stat(si->what, &sb) == 0) == wanted_state) {
+ job_log(j, LOG_DEBUG, "KeepAlive: The following path %s: %s", wanted_state ? "exists" : "is missing", si->what);
+ return true;
+ } else {
+ if (wanted_state) { /* File is not there but we wish it was. */
+ if (si->fd != -1 && !si->watching_parent) { /* Need to be watching the parent now. */
+ (void)job_assumes(j, runtime_close(si->fd) == 0);
+ si->fd = -1;
+ semaphoreitem_watch(j, si);
+ }
+ } else { /* File is there but we wish it wasn't. */
+ if (si->fd != -1 && si->watching_parent) { /* Need to watch the file now. */
+ (void)job_assumes(j, runtime_close(si->fd) == 0);
+ si->fd = -1;
+ semaphoreitem_watch(j, si);
+ }
+ }
+ }
+ break;
+ case PATH_CHANGES:
+ break;
+ case DIR_NOT_EMPTY:
+ if (-1 == (qdir_file_cnt = dir_has_files(j, si->what))) {
+ job_log_error(j, LOG_ERR, "Failed to count the number of files in \"%s\"", si->what);
+ } else if (qdir_file_cnt > 0) {
+ job_log(j, LOG_DEBUG, "KeepAlive: Directory is not empty: %s", si->what);
+ return true;
+ }
+ break;
+ }
+ }
+
+ return false;
+}
+
+const char *
+job_active(job_t j)
+{
+ struct machservice *ms;
+ if (j->p && j->shutdown_monitor) {
+ return "Monitoring shutdown";
+ }
+ if (j->p) {
+ return "PID is still valid";
+ }
+
+ if (j->mgr->shutting_down && j->log_redirect_fd) {
+ (void)job_assumes(j, runtime_close(j->log_redirect_fd) != -1);
+ j->log_redirect_fd = 0;
+ }
+
+ if (j->log_redirect_fd) {
+ if (job_assumes(j, j->legacy_LS_job)) {
+ return "Standard out/error is still valid";
+ } else {
+ (void)job_assumes(j, runtime_close(j->log_redirect_fd) != -1);
+ j->log_redirect_fd = 0;
+ }
+ }
+
+ if (j->priv_port_has_senders) {
+ return "Privileged Port still has outstanding senders";
+ }
+
+ SLIST_FOREACH(ms, &j->machservices, sle) {
+ if (ms->recv && machservice_active(ms)) {
+ return "Mach service is still active";
+ }
+ }
+
+ return NULL;
+}
+
+void
+machservice_watch(job_t j, struct machservice *ms)
+{
+ if (ms->recv) {
+ (void)job_assumes(j, runtime_add_mport(ms->port, NULL, 0) == KERN_SUCCESS);
+ }
+}
+
+void
+machservice_ignore(job_t j, struct machservice *ms)
+{
+ (void)job_assumes(j, runtime_remove_mport(ms->port) == KERN_SUCCESS);
+}
+
+void
+machservice_resetport(job_t j, struct machservice *ms)
+{
+ LIST_REMOVE(ms, port_hash_sle);
+ (void)job_assumes(j, launchd_mport_close_recv(ms->port) == KERN_SUCCESS);
+ (void)job_assumes(j, launchd_mport_deallocate(ms->port) == KERN_SUCCESS);
+ ms->gen_num++;
+ (void)job_assumes(j, launchd_mport_create_recv(&ms->port) == KERN_SUCCESS);
+ (void)job_assumes(j, launchd_mport_make_send(ms->port) == KERN_SUCCESS);
+ LIST_INSERT_HEAD(&port_hash[HASH_PORT(ms->port)], ms, port_hash_sle);
+}
+
+struct machservice *
+machservice_new(job_t j, const char *name, mach_port_t *serviceport, bool pid_local)
+{
+ struct machservice *ms = calloc(1, sizeof(struct machservice) + strlen(name) + 1);
+
+ if (!job_assumes(j, ms != NULL)) {
+ return NULL;
+ }
+
+ strcpy((char *)ms->name, name);
+ ms->job = j;
+ ms->gen_num = 1;
+ ms->per_pid = pid_local;
+
+ if (likely(*serviceport == MACH_PORT_NULL)) {
+ if (!job_assumes(j, launchd_mport_create_recv(&ms->port) == KERN_SUCCESS)) {
+ goto out_bad;
+ }
+
+ if (!job_assumes(j, launchd_mport_make_send(ms->port) == KERN_SUCCESS)) {
+ goto out_bad2;
+ }
+ *serviceport = ms->port;
+ ms->recv = true;
+ } else {
+ ms->port = *serviceport;
+ ms->isActive = true;
+ }
+
+ SLIST_INSERT_HEAD(&j->machservices, ms, sle);
+
+ jobmgr_t where2put = j->mgr;
+ /* XPC domains are separate from Mach bootstraps. */
+ if (!(j->mgr->properties & BOOTSTRAP_PROPERTY_XPC_DOMAIN)) {
+ if (g_flat_mach_namespace && !(j->mgr->properties & BOOTSTRAP_PROPERTY_EXPLICITSUBSET)) {
+ where2put = root_jobmgr;
+ }
+ }
+
+ /* Don't allow MachServices added by multiple-instance jobs to be looked up by others.
+ * We could just do this with a simple bit, but then we'd have to uniquify the
+ * names ourselves to avoid collisions. This is just easier.
+ */
+ if (!j->dedicated_instance) {
+ LIST_INSERT_HEAD(&where2put->ms_hash[hash_ms(ms->name)], ms, name_hash_sle);
+ }
+ LIST_INSERT_HEAD(&port_hash[HASH_PORT(ms->port)], ms, port_hash_sle);
+
+ job_log(j, LOG_DEBUG, "Mach service added%s: %s", (j->mgr->properties & BOOTSTRAP_PROPERTY_EXPLICITSUBSET) ? " to private namespace" : "", name);
+
+ return ms;
+out_bad2:
+ (void)job_assumes(j, launchd_mport_close_recv(ms->port) == KERN_SUCCESS);
+out_bad:
+ free(ms);
+ return NULL;
+}
+
+#ifndef __LAUNCH_DISABLE_XPC_SUPPORT__
+struct machservice *
+machservice_new_alias(job_t j, struct machservice *orig)
+{
+ struct machservice *ms = calloc(1, sizeof(struct machservice) + strlen(orig->name) + 1);
+ if (job_assumes(j, ms != NULL)) {
+ strcpy((char *)ms->name, orig->name);
+ ms->alias = orig;
+ ms->job = j;
+
+ LIST_INSERT_HEAD(&j->mgr->ms_hash[hash_ms(ms->name)], ms, name_hash_sle);
+ SLIST_INSERT_HEAD(&j->machservices, ms, sle);
+ jobmgr_log(j->mgr, LOG_DEBUG, "Service aliased into job manager: %s", orig->name);
+ }
+
+ return ms;
+}
+#endif
+
+bootstrap_status_t
+machservice_status(struct machservice *ms)
+{
+ ms = ms->alias ? ms->alias : ms;
+ if (ms->isActive) {
+ return BOOTSTRAP_STATUS_ACTIVE;
+ } else if (ms->job->ondemand) {
+ return BOOTSTRAP_STATUS_ON_DEMAND;
+ } else {
+ return BOOTSTRAP_STATUS_INACTIVE;
+ }
+}
+
+void
+job_setup_exception_port(job_t j, task_t target_task)
+{
+ struct machservice *ms;
+ thread_state_flavor_t f = 0;
+ mach_port_t exc_port = the_exception_server;
+
+ if (unlikely(j->alt_exc_handler)) {
+ ms = jobmgr_lookup_service(j->mgr, j->alt_exc_handler, true, 0);
+ if (likely(ms)) {
+ exc_port = machservice_port(ms);
+ } else {
+ job_log(j, LOG_WARNING, "Falling back to default Mach exception handler. Could not find: %s", j->alt_exc_handler);
+ }
+ } else if (unlikely(j->internal_exc_handler)) {
+ exc_port = runtime_get_kernel_port();
+ } else if (unlikely(!exc_port)) {
+ return;
+ }
+
+#if defined (__ppc__) || defined(__ppc64__)
+ f = PPC_THREAD_STATE64;
+#elif defined(__i386__) || defined(__x86_64__)
+ f = x86_THREAD_STATE;
+#elif defined(__arm__)
+ f = ARM_THREAD_STATE;
+#else
+#error "unknown architecture"
+#endif
+
+ if (likely(target_task)) {
+ (void)job_assumes(j, task_set_exception_ports(target_task, EXC_MASK_CRASH, exc_port, EXCEPTION_STATE_IDENTITY | MACH_EXCEPTION_CODES, f) == KERN_SUCCESS);
+ } else if (pid1_magic && the_exception_server) {
+ mach_port_t mhp = mach_host_self();
+ (void)job_assumes(j, host_set_exception_ports(mhp, EXC_MASK_CRASH, the_exception_server, EXCEPTION_STATE_IDENTITY | MACH_EXCEPTION_CODES, f) == KERN_SUCCESS);
+ job_assumes(j, launchd_mport_deallocate(mhp) == KERN_SUCCESS);
+ }
+}
+
+void
+job_set_exception_port(job_t j, mach_port_t port)
+{
+ if (unlikely(!the_exception_server)) {
+ the_exception_server = port;
+ job_setup_exception_port(j, 0);
+ } else {
+ job_log(j, LOG_WARNING, "The exception server is already claimed!");
+ }
+}
+
+void
+machservice_setup_options(launch_data_t obj, const char *key, void *context)
+{
+ struct machservice *ms = context;
+ mach_port_t mhp = mach_host_self();
+ int which_port;
+ bool b;
+
+ if (!job_assumes(ms->job, mhp != MACH_PORT_NULL)) {
+ return;
+ }
+
+ switch (launch_data_get_type(obj)) {
case LAUNCH_DATA_INTEGER:
- which_port = launch_data_get_integer(obj);
+ which_port = (int)launch_data_get_integer(obj); /* XXX we should bound check this... */
if (strcasecmp(key, LAUNCH_JOBKEY_MACH_TASKSPECIALPORT) == 0) {
- launchd_assumes((errno = task_set_special_port(mts, which_port, ms->port)) == KERN_SUCCESS);
- } else if (strcasecmp(key, LAUNCH_JOBKEY_MACH_HOSTSPECIALPORT) == 0 && getpid() == 1) {
- launchd_assumes((errno = host_set_special_port(mhp, which_port, ms->port)) == KERN_SUCCESS);
+ switch (which_port) {
+ case TASK_KERNEL_PORT:
+ case TASK_HOST_PORT:
+ case TASK_NAME_PORT:
+ case TASK_BOOTSTRAP_PORT:
+ /* I find it a little odd that zero isn't reserved in the header.
+ * Normally Mach is fairly good about this convention... */
+ case 0:
+ job_log(ms->job, LOG_WARNING, "Tried to set a reserved task special port: %d", which_port);
+ break;
+ default:
+ ms->special_port_num = which_port;
+ SLIST_INSERT_HEAD(&special_ports, ms, special_port_sle);
+ break;
+ }
+ } else if (strcasecmp(key, LAUNCH_JOBKEY_MACH_HOSTSPECIALPORT) == 0 && pid1_magic) {
+ if (which_port > HOST_MAX_SPECIAL_KERNEL_PORT) {
+ (void)job_assumes(ms->job, (errno = host_set_special_port(mhp, which_port, ms->port)) == KERN_SUCCESS);
+ } else {
+ job_log(ms->job, LOG_WARNING, "Tried to set a reserved host special port: %d", which_port);
+ }
+ }
+ case LAUNCH_DATA_BOOL:
+ b = launch_data_get_bool(obj);
+ if (strcasecmp(key, LAUNCH_JOBKEY_MACH_ENTERKERNELDEBUGGERONCLOSE) == 0) {
+ ms->debug_on_close = b;
+ } else if (strcasecmp(key, LAUNCH_JOBKEY_MACH_RESETATCLOSE) == 0) {
+ ms->reset = b;
+ } else if (strcasecmp(key, LAUNCH_JOBKEY_MACH_HIDEUNTILCHECKIN) == 0) {
+ ms->hide = b;
+ } else if (strcasecmp(key, LAUNCH_JOBKEY_MACH_EXCEPTIONSERVER) == 0) {
+ job_set_exception_port(ms->job, ms->port);
+ } else if (strcasecmp(key, LAUNCH_JOBKEY_MACH_KUNCSERVER) == 0) {
+ ms->kUNCServer = b;
+ (void)job_assumes(ms->job, host_set_UNDServer(mhp, ms->port) == KERN_SUCCESS);
+ } else if (strcasecmp(key, LAUNCH_JOBKEY_MACH_PINGEVENTUPDATES) == 0) {
+ ms->event_update_port = b;
+ }
+ break;
+ case LAUNCH_DATA_STRING:
+ if (strcasecmp(key, LAUNCH_JOBKEY_MACH_DRAINMESSAGESONCRASH) == 0) {
+ const char *option = launch_data_get_string(obj);
+ if (strcasecmp(option, "One") == 0) {
+ ms->drain_one_on_crash = true;
+ } else if (strcasecmp(option, "All") == 0) {
+ ms->drain_all_on_crash = true;
+ }
+ }
+ break;
+ case LAUNCH_DATA_DICTIONARY:
+ job_set_exception_port(ms->job, ms->port);
+ break;
+ default:
+ break;
+ }
+
+ job_assumes(ms->job, launchd_mport_deallocate(mhp) == KERN_SUCCESS);
+}
+
+void
+machservice_setup(launch_data_t obj, const char *key, void *context)
+{
+ job_t j = context;
+ struct machservice *ms;
+ mach_port_t p = MACH_PORT_NULL;
+
+ if (unlikely(ms = jobmgr_lookup_service(j->mgr, key, false, 0))) {
+ job_log(j, LOG_WARNING, "Conflict with job: %s over Mach service: %s", ms->job->label, key);
+ return;
+ }
+
+ if (!job_assumes(j, (ms = machservice_new(j, key, &p, false)) != NULL)) {
+ return;
+ }
+
+ ms->isActive = false;
+ ms->upfront = true;
+
+ if (launch_data_get_type(obj) == LAUNCH_DATA_DICTIONARY) {
+ launch_data_dict_iterate(obj, machservice_setup_options, ms);
+ }
+}
+
+jobmgr_t
+jobmgr_do_garbage_collection(jobmgr_t jm)
+{
+ jobmgr_t jmi = NULL, jmn = NULL;
+ SLIST_FOREACH_SAFE(jmi, &jm->submgrs, sle, jmn) {
+ jobmgr_do_garbage_collection(jmi);
+ }
+
+ if (!jm->shutting_down) {
+ return jm;
+ }
+
+ if (SLIST_EMPTY(&jm->submgrs)) {
+ jobmgr_log(jm, LOG_DEBUG, "No submanagers left.");
+ } else {
+ jobmgr_log(jm, LOG_DEBUG, "Still have submanagers.");
+ }
+
+ size_t actives = 0;
+ job_t ji = NULL, jn = NULL;
+ LIST_FOREACH_SAFE(ji, &jm->jobs, sle, jn) {
+ if (ji->anonymous) {
+ continue;
+ }
+
+ /* Let the shutdown monitor be up until the very end. */
+ if (ji->shutdown_monitor) {
+ continue;
+ }
+
+ /* On our first pass through, open a transaction for all the jobs that
+ * need to be dirty at shutdown. We'll close these transactions once the
+ * jobs that do not need to be dirty at shutdown have all exited.
+ */
+ if (ji->dirty_at_shutdown && !jm->shutdown_jobs_dirtied) {
+ job_open_shutdown_transaction(ji);
+ }
+
+ const char *active = job_active(ji);
+ if (!active) {
+ job_remove(ji);
+ } else {
+ job_log(ji, LOG_DEBUG, "Job is active: %s", active);
+ job_stop(ji);
+
+ if (ji->p && !ji->dirty_at_shutdown) {
+ /* We really only care if the job has not yet been reaped.
+ * There's no reason to delay shutdown if a Mach port has not
+ * yet been sent back to us, for example. While we're shutting
+ * all the "normal" jobs down, do not count the
+ * dirty-at-shutdown jobs toward the total of actives.
+ *
+ * Note that there's a potential race here where we may not get
+ * a port back in time, so that when we hit jobmgr_remove(), we
+ * end up removing the job and then our attempt to close the
+ * Mach port will fail. But at that point, the failure won't
+ * even make it to the syslog, so not a big deal.
+ */
+ actives++;
+ }
+
+ if (ji->clean_kill) {
+ job_log(ji, LOG_DEBUG, "Job was killed cleanly.");
+ } else {
+ job_log(ji, LOG_DEBUG, "Job was sent SIGTERM%s.", ji->sent_sigkill ? " and SIGKILL" : "");
+ }
+ }
+ }
+
+ jm->shutdown_jobs_dirtied = true;
+ if (actives == 0) {
+ if (!jm->shutdown_jobs_cleaned) {
+ LIST_FOREACH(ji, &jm->jobs, sle) {
+ if (!ji->anonymous) {
+ job_close_shutdown_transaction(ji);
+ actives++;
+ }
+ }
+
+ jm->shutdown_jobs_cleaned = true;
+ } else if (jm->monitor_shutdown && _s_shutdown_monitor) {
+ /* The rest of shutdown has completed, so we can kill the shutdown
+ * monitor now like it was any other job.
+ */
+ _s_shutdown_monitor->shutdown_monitor = false;
+ actives = 1;
+
+ job_log(_s_shutdown_monitor, LOG_NOTICE | LOG_CONSOLE, "Stopping shutdown monitor.");
+ job_stop(_s_shutdown_monitor);
+ _s_shutdown_monitor = NULL;
+ }
+ }
+
+ jobmgr_t r = jm;
+ if (SLIST_EMPTY(&jm->submgrs) && actives == 0) {
+ jobmgr_log(jm, LOG_DEBUG, "Removing.");
+ jobmgr_remove(jm);
+ r = NULL;
+ }
+
+ return r;
+}
+
+void
+jobmgr_kill_stray_children(jobmgr_t jm, pid_t *p, size_t np)
+{
+ /* I maintain that stray processes should be at the mercy of launchd during shutdown,
+ * but nevertheless, things like diskimages-helper can stick around, and SIGKILLing
+ * them can result in data loss. So we send SIGTERM to all the strays and don't wait
+ * for them to exit before moving on.
+ *
+ * See rdar://problem/6562592
+ */
+ size_t i = 0;
+ for (i = 0; i < np; i++) {
+ if (p[i] != 0) {
+ jobmgr_log(jm, LOG_DEBUG | LOG_CONSOLE, "Sending SIGTERM to PID %u and continuing...", p[i]);
+ (void)jobmgr_assumes(jm, runtime_kill(p[i], SIGTERM) != -1);
+ }
+ }
+}
+
+void
+jobmgr_log_stray_children(jobmgr_t jm, bool kill_strays)
+{
+ size_t kp_skipped = 0, len = sizeof(pid_t) * get_kern_max_proc();
+ pid_t *pids = NULL;
+ int i = 0, kp_cnt = 0;
+
+ if (likely(jm->parentmgr || !pid1_magic)) {
+ return;
+ }
+
+ if (!jobmgr_assumes(jm, (pids = malloc(len)) != NULL)) {
+ return;
+ }
+
+ runtime_ktrace0(RTKT_LAUNCHD_FINDING_ALL_STRAYS);
+
+ if (!jobmgr_assumes(jm, (kp_cnt = proc_listallpids(pids, len)) != -1)) {
+ goto out;
+ }
+
+ pid_t *ps = (pid_t *)calloc(sizeof(pid_t), kp_cnt);
+ for (i = 0; i < kp_cnt; i++) {
+ struct proc_bsdshortinfo proc;
+ if (proc_pidinfo(pids[i], PROC_PIDT_SHORTBSDINFO, 1, &proc, PROC_PIDT_SHORTBSDINFO_SIZE) == 0) {
+ if (errno != ESRCH) {
+ jobmgr_assumes(jm, errno == 0);
+ }
+
+ kp_skipped++;
+ continue;
+ }
+
+ pid_t p_i = pids[i];
+ pid_t pp_i = proc.pbsi_ppid;
+ pid_t pg_i = proc.pbsi_pgid;
+ const char *z = (proc.pbsi_status == SZOMB) ? "zombie " : "";
+ const char *n = proc.pbsi_comm;
+
+ if (unlikely(p_i == 0 || p_i == 1)) {
+ kp_skipped++;
+ continue;
+ }
+
+ if (_s_shutdown_monitor && pp_i == _s_shutdown_monitor->p) {
+ kp_skipped++;
+ continue;
+ }
+
+ /* We might have some jobs hanging around that we've decided to shut down in spite of. */
+ job_t j = jobmgr_find_by_pid(jm, p_i, false);
+ if (!j || (j && j->anonymous)) {
+ jobmgr_log(jm, LOG_INFO | LOG_CONSOLE, "Stray %s%s at shutdown: PID %u PPID %u PGID %u %s", z, j ? "anonymous job" : "process", p_i, pp_i, pg_i, n);
+
+ int status = 0;
+ if (pp_i == getpid() && !jobmgr_assumes(jm, proc.pbsi_status != SZOMB)) {
+ if (jobmgr_assumes(jm, waitpid(p_i, &status, WNOHANG) == 0)) {
+ jobmgr_log(jm, LOG_INFO | LOG_CONSOLE, "Unreaped zombie stray exited with status %i.", WEXITSTATUS(status));
+ }
+ kp_skipped++;
+ } else {
+ job_t leader = jobmgr_find_by_pid(jm, pg_i, false);
+ /* See rdar://problem/6745714. Some jobs have child processes that back kernel state,
+ * so we don't want to terminate them. Long-term, I'd really like to provide shutdown
+ * hints to the kernel along the way, so that it could shutdown certain subsystems when
+ * their userspace emissaries go away, before the call to reboot(2).
+ */
+ if (leader && leader->ignore_pg_at_shutdown) {
+ kp_skipped++;
+ } else {
+ ps[i] = p_i;
+ }
+ }
+ } else {
+ kp_skipped++;
+ }
+ }
+
+ if ((kp_cnt - kp_skipped > 0) && kill_strays) {
+ jobmgr_kill_stray_children(jm, ps, kp_cnt - kp_skipped);
+ }
+
+ free(ps);
+out:
+ free(pids);
+}
+
+jobmgr_t
+jobmgr_parent(jobmgr_t jm)
+{
+ return jm->parentmgr;
+}
+
+void
+job_uncork_fork(job_t j)
+{
+ pid_t c = j->p;
+
+ job_log(j, LOG_DEBUG, "Uncorking the fork().");
+ /* this unblocks the child and avoids a race
+ * between the above fork() and the kevent_mod() */
+ (void)job_assumes(j, write(j->fork_fd, &c, sizeof(c)) == sizeof(c));
+ (void)job_assumes(j, runtime_close(j->fork_fd) != -1);
+ j->fork_fd = 0;
+}
+
+jobmgr_t
+jobmgr_new(jobmgr_t jm, mach_port_t requestorport, mach_port_t transfer_port, bool sflag, const char *name, bool skip_init, mach_port_t asport)
+{
+ mach_msg_size_t mxmsgsz;
+ job_t bootstrapper = NULL;
+ jobmgr_t jmr;
+
+ launchd_assert(offsetof(struct jobmgr_s, kqjobmgr_callback) == 0);
+
+ if (unlikely(jm && requestorport == MACH_PORT_NULL)) {
+ jobmgr_log(jm, LOG_ERR, "Mach sub-bootstrap create request requires a requester port");
+ return NULL;
+ }
+
+ jmr = calloc(1, sizeof(struct jobmgr_s) + (name ? (strlen(name) + 1) : NAME_MAX + 1));
+
+ if (!jobmgr_assumes(jm, jmr != NULL)) {
+ return NULL;
+ }
+
+ if (jm == NULL) {
+ root_jobmgr = jmr;
+ }
+
+ jmr->kqjobmgr_callback = jobmgr_callback;
+ strcpy(jmr->name_init, name ? name : "Under construction");
+
+ jmr->req_port = requestorport;
+
+ if ((jmr->parentmgr = jm)) {
+ SLIST_INSERT_HEAD(&jm->submgrs, jmr, sle);
+ }
+
+ if (jm && !jobmgr_assumes(jmr, launchd_mport_notify_req(jmr->req_port, MACH_NOTIFY_DEAD_NAME) == KERN_SUCCESS)) {
+ goto out_bad;
+ }
+
+ if (transfer_port != MACH_PORT_NULL) {
+ (void)jobmgr_assumes(jmr, jm != NULL);
+ jmr->jm_port = transfer_port;
+ } else if (!jm && !pid1_magic) {
+ char *trusted_fd = getenv(LAUNCHD_TRUSTED_FD_ENV);
+ name_t service_buf;
+
+ snprintf(service_buf, sizeof(service_buf), "com.apple.launchd.peruser.%u", getuid());
+
+ if (!jobmgr_assumes(jmr, bootstrap_check_in(bootstrap_port, service_buf, &jmr->jm_port) == 0)) {
+ goto out_bad;
+ }
+
+ if (trusted_fd) {
+ int dfd, lfd = (int) strtol(trusted_fd, NULL, 10);
+
+ if ((dfd = dup(lfd)) >= 0) {
+ (void)jobmgr_assumes(jmr, runtime_close(dfd) != -1);
+ (void)jobmgr_assumes(jmr, runtime_close(lfd) != -1);
+ }
+
+ unsetenv(LAUNCHD_TRUSTED_FD_ENV);
+ }
+
+ /* cut off the Libc cache, we don't want to deadlock against ourself */
+ inherited_bootstrap_port = bootstrap_port;
+ bootstrap_port = MACH_PORT_NULL;
+ launchd_assert(launchd_mport_notify_req(inherited_bootstrap_port, MACH_NOTIFY_DEAD_NAME) == KERN_SUCCESS);
+
+ /* We set this explicitly as we start each child */
+ launchd_assert(launchd_set_bport(MACH_PORT_NULL) == KERN_SUCCESS);
+ } else if (!jobmgr_assumes(jmr, launchd_mport_create_recv(&jmr->jm_port) == KERN_SUCCESS)) {
+ goto out_bad;
+ }
+
+ if (!name) {
+ sprintf(jmr->name_init, "%u", MACH_PORT_INDEX(jmr->jm_port));
+ }
+
+ /* Sigh... at the moment, MIG has maxsize == sizeof(reply union) */
+ mxmsgsz = (typeof(mxmsgsz)) sizeof(union __RequestUnion__job_mig_protocol_vproc_subsystem);
+ if (job_mig_protocol_vproc_subsystem.maxsize > mxmsgsz) {
+ mxmsgsz = job_mig_protocol_vproc_subsystem.maxsize;
+ }
+
+ /* Total hacks. But the MIG server loop is too generic, and the more dynamic
+ * parts of it haven't been tested, or if they have, it was a very long time
+ * ago.
+ */
+ if (xpc_events_xpc_events_subsystem.maxsize > mxmsgsz) {
+ mxmsgsz = xpc_events_xpc_events_subsystem.maxsize;
+ }
+ if (xpc_domain_xpc_domain_subsystem.maxsize > mxmsgsz) {
+ mxmsgsz = xpc_domain_xpc_domain_subsystem.maxsize;
+ }
+
+ if (!jm) {
+ (void)jobmgr_assumes(jmr, kevent_mod(SIGTERM, EVFILT_SIGNAL, EV_ADD, 0, 0, jmr) != -1);
+ (void)jobmgr_assumes(jmr, kevent_mod(SIGUSR1, EVFILT_SIGNAL, EV_ADD, 0, 0, jmr) != -1);
+ (void)jobmgr_assumes(jmr, kevent_mod(SIGUSR2, EVFILT_SIGNAL, EV_ADD, 0, 0, jmr) != -1);
+ (void)jobmgr_assumes(jmr, kevent_mod(0, EVFILT_FS, EV_ADD, VQ_MOUNT|VQ_UNMOUNT|VQ_UPDATE, 0, jmr) != -1);
+ }
+
+ if (name && !skip_init) {
+ bootstrapper = jobmgr_init_session(jmr, name, sflag);
+ }
+
+ if (!bootstrapper || !bootstrapper->weird_bootstrap) {
+ if (!jobmgr_assumes(jmr, runtime_add_mport(jmr->jm_port, protocol_vproc_server, mxmsgsz) == KERN_SUCCESS)) {
+ goto out_bad;
+ }
+ }
+
+ jobmgr_log(jmr, LOG_DEBUG, "Created job manager%s%s", jm ? " with parent: " : ".", jm ? jm->name : "");
+
+ if (bootstrapper) {
+ bootstrapper->asport = asport;
+
+ jobmgr_log(jmr, LOG_DEBUG, "Bootstrapping new job manager with audit session %u", asport);
+ (void)jobmgr_assumes(jmr, job_dispatch(bootstrapper, true) != NULL);
+ } else {
+ jmr->req_asport = asport;
+ }
+
+ if (asport != MACH_PORT_NULL) {
+ (void)jobmgr_assumes(jmr, launchd_mport_copy_send(asport) == KERN_SUCCESS);
+ }
+
+ if (jmr->parentmgr) {
+ runtime_add_weak_ref();
+ }
+
+ return jmr;
+
+out_bad:
+ if (jmr) {
+ jobmgr_remove(jmr);
+ if (jm == NULL) {
+ root_jobmgr = NULL;
+ }
+ }
+ return NULL;
+}
+
+#ifndef __LAUNCH_DISABLE_XPC_SUPPORT__
+jobmgr_t
+jobmgr_new_xpc_singleton_domain(jobmgr_t jm, name_t name)
+{
+ jobmgr_t new = NULL;
+
+ /* These job managers are basically singletons, so we use the root Mach
+ * bootstrap port as their requestor ports so they'll never go away.
+ */
+ mach_port_t req_port = root_jobmgr->jm_port;
+ if (jobmgr_assumes(jm, launchd_mport_make_send(req_port) == KERN_SUCCESS)) {
+ new = jobmgr_new(root_jobmgr, req_port, MACH_PORT_NULL, false, name, true, MACH_PORT_NULL);
+ if (new) {
+ new->properties |= BOOTSTRAP_PROPERTY_XPC_SINGLETON;
+ new->properties |= BOOTSTRAP_PROPERTY_XPC_DOMAIN;
+ new->xpc_singleton = true;
+ }
+ }
+
+ return new;
+}
+
+jobmgr_t
+jobmgr_find_xpc_per_user_domain(jobmgr_t jm, uid_t uid)
+{
+ jobmgr_t jmi = NULL;
+ LIST_FOREACH(jmi, &_s_xpc_user_domains, xpc_le) {
+ if (jmi->req_euid == uid) {
+ return jmi;
+ }
+ }
+
+ name_t name;
+ (void)snprintf(name, sizeof(name), "com.apple.xpc.domain.peruser.%u", uid);
+ jmi = jobmgr_new_xpc_singleton_domain(jm, name);
+ if (jobmgr_assumes(jm, jmi != NULL)) {
+ /* We need to create a per-user launchd for this UID if there isn't one
+ * already so we can grab the bootstrap port.
+ */
+ job_t puj = jobmgr_lookup_per_user_context_internal(NULL, uid, &jmi->req_bsport);
+ if (jobmgr_assumes(jmi, puj != NULL)) {
+ (void)jobmgr_assumes(jmi, launchd_mport_copy_send(puj->asport) == KERN_SUCCESS);
+ (void)jobmgr_assumes(jmi, launchd_mport_copy_send(jmi->req_bsport) == KERN_SUCCESS);
+ jmi->shortdesc = "per-user";
+ jmi->req_asport = puj->asport;
+ jmi->req_asid = puj->asid;
+ jmi->req_euid = uid;
+ jmi->req_egid = -1;
+
+ LIST_INSERT_HEAD(&_s_xpc_user_domains, jmi, xpc_le);
+ } else {
+ jobmgr_remove(jmi);
+ }
+ }
+
+ return jmi;
+}
+
+jobmgr_t
+jobmgr_find_xpc_per_session_domain(jobmgr_t jm, au_asid_t asid)
+{
+ jobmgr_t jmi = NULL;
+ LIST_FOREACH(jmi, &_s_xpc_session_domains, xpc_le) {
+ if (jmi->req_asid == asid) {
+ return jmi;
+ }
+ }
+
+ name_t name;
+ (void)snprintf(name, sizeof(name), "com.apple.xpc.domain.persession.%i", asid);
+ jmi = jobmgr_new_xpc_singleton_domain(jm, name);
+ if (jobmgr_assumes(jm, jmi != NULL)) {
+ (void)jobmgr_assumes(jmi, launchd_mport_make_send(root_jobmgr->jm_port) == KERN_SUCCESS);
+ jmi->shortdesc = "per-session";
+ jmi->req_bsport = root_jobmgr->jm_port;
+ (void)jobmgr_assumes(jmi, audit_session_port(asid, &jmi->req_asport) == 0);
+ jmi->req_asid = asid;
+ jmi->req_euid = -1;
+ jmi->req_egid = -1;
+
+ LIST_INSERT_HEAD(&_s_xpc_session_domains, jmi, xpc_le);
+ } else {
+ jobmgr_remove(jmi);
+ }
+
+ return jmi;
+}
+#endif
+
+job_t
+jobmgr_init_session(jobmgr_t jm, const char *session_type, bool sflag)
+{
+ const char *bootstrap_tool[] = { "/bin/launchctl", "bootstrap", "-S", session_type, sflag ? "-s" : NULL, NULL };
+ char thelabel[1000];
+ job_t bootstrapper;
+
+ snprintf(thelabel, sizeof(thelabel), "com.apple.launchctl.%s", session_type);
+ bootstrapper = job_new(jm, thelabel, NULL, bootstrap_tool);
+
+ if (jobmgr_assumes(jm, bootstrapper != NULL) && (jm->parentmgr || !pid1_magic)) {
+ bootstrapper->is_bootstrapper = true;
+ char buf[100];
+
+ /* <rdar://problem/5042202> launchd-201: can't ssh in with AFP OD account (hangs) */
+ snprintf(buf, sizeof(buf), "0x%X:0:0", getuid());
+ envitem_new(bootstrapper, "__CF_USER_TEXT_ENCODING", buf, false, false);
+ bootstrapper->weird_bootstrap = true;
+ (void)jobmgr_assumes(jm, job_setup_machport(bootstrapper));
+ } else if (bootstrapper && strncmp(session_type, VPROCMGR_SESSION_SYSTEM, sizeof(VPROCMGR_SESSION_SYSTEM)) == 0) {
+ bootstrapper->is_bootstrapper = true;
+ if (jobmgr_assumes(jm, pid1_magic)) {
+ /* Have our system bootstrapper print out to the console. */
+ bootstrapper->stdoutpath = strdup(_PATH_CONSOLE);
+ bootstrapper->stderrpath = strdup(_PATH_CONSOLE);
+
+ if (g_console) {
+ (void)jobmgr_assumes(jm, kevent_mod((uintptr_t)fileno(g_console), EVFILT_VNODE, EV_ADD | EV_ONESHOT, NOTE_REVOKE, 0, jm) != -1);
+ }
+ }
+ }
+
+ jm->session_initialized = true;
+ return bootstrapper;
+}
+
+jobmgr_t
+jobmgr_delete_anything_with_port(jobmgr_t jm, mach_port_t port)
+{
+ struct machservice *ms, *next_ms;
+ jobmgr_t jmi, jmn;
+
+ /* Mach ports, unlike Unix descriptors, are reference counted. In other
+ * words, when some program hands us a second or subsequent send right
+ * to a port we already have open, the Mach kernel gives us the same
+ * port number back and increments an reference count associated with
+ * the port. This forces us, when discovering that a receive right at
+ * the other end has been deleted, to wander all of our objects to see
+ * what weird places clients might have handed us the same send right
+ * to use.
+ */
+
+ if (jm == root_jobmgr) {
+ if (port == inherited_bootstrap_port) {
+ (void)jobmgr_assumes(jm, launchd_mport_deallocate(port) == KERN_SUCCESS);
+ inherited_bootstrap_port = MACH_PORT_NULL;
+
+ return jobmgr_shutdown(jm);
+ }
+
+ LIST_FOREACH_SAFE(ms, &port_hash[HASH_PORT(port)], port_hash_sle, next_ms) {
+ if (ms->port == port && !ms->recv) {
+ machservice_delete(ms->job, ms, true);
+ }
+ }
+ }
+
+ SLIST_FOREACH_SAFE(jmi, &jm->submgrs, sle, jmn) {
+ jobmgr_delete_anything_with_port(jmi, port);
+ }
+
+ if (jm->req_port == port) {
+ jobmgr_log(jm, LOG_DEBUG, "Request port died: %i", MACH_PORT_INDEX(port));
+ return jobmgr_shutdown(jm);
+ }
+
+ return jm;
+}
+
+struct machservice *
+jobmgr_lookup_service(jobmgr_t jm, const char *name, bool check_parent, pid_t target_pid)
+{
+ struct machservice *ms;
+ job_t target_j;
+
+ jobmgr_log(jm, LOG_DEBUG, "Looking up %sservice %s", target_pid ? "per-PID " : "", name);
+
+ if (target_pid) {
+ /* This is a hack to let FileSyncAgent look up per-PID Mach services from the Background
+ * bootstrap in other bootstraps.
+ */
+
+ /* Start in the given bootstrap. */
+ if (unlikely((target_j = jobmgr_find_by_pid(jm, target_pid, false)) == NULL)) {
+ /* If we fail, do a deep traversal. */
+ if (unlikely((target_j = jobmgr_find_by_pid_deep(root_jobmgr, target_pid, true)) == NULL)) {
+ jobmgr_log(jm, LOG_DEBUG, "Didn't find PID %i", target_pid);
+ return NULL;
+ }
+ }
+
+ SLIST_FOREACH(ms, &target_j->machservices, sle) {
+ if (ms->per_pid && strcmp(name, ms->name) == 0) {
+ return ms;
+ }
+ }
+
+ job_log(target_j, LOG_DEBUG, "Didn't find per-PID Mach service: %s", name);
+ return NULL;
+ }
+
+ jobmgr_t where2look = jm;
+ /* XPC domains are separate from Mach bootstraps. */
+ if (!(jm->properties & BOOTSTRAP_PROPERTY_XPC_DOMAIN)) {
+ if (g_flat_mach_namespace && !(jm->properties & BOOTSTRAP_PROPERTY_EXPLICITSUBSET)) {
+ where2look = root_jobmgr;
+ }
+ }
+
+ LIST_FOREACH(ms, &where2look->ms_hash[hash_ms(name)], name_hash_sle) {
+ if (!ms->per_pid && strcmp(name, ms->name) == 0) {
+ return ms;
+ }
+ }
+
+ if (jm->parentmgr == NULL || !check_parent) {
+ return NULL;
+ }
+
+ return jobmgr_lookup_service(jm->parentmgr, name, true, 0);
+}
+
+mach_port_t
+machservice_port(struct machservice *ms)
+{
+ return ms->port;
+}
+
+job_t
+machservice_job(struct machservice *ms)
+{
+ return ms->job;
+}
+
+bool
+machservice_hidden(struct machservice *ms)
+{
+ return ms->hide;
+}
+
+bool
+machservice_active(struct machservice *ms)
+{
+ return ms->isActive;
+}
+
+const char *
+machservice_name(struct machservice *ms)
+{
+ return ms->name;
+}
+
+void
+machservice_drain_port(struct machservice *ms)
+{
+ bool drain_one = ms->drain_one_on_crash;
+ bool drain_all = ms->drain_all_on_crash;
+
+ if (!job_assumes(ms->job, (drain_one || drain_all) == true)) {
+ return;
+ }
+
+ job_log(ms->job, LOG_INFO, "Draining %s...", ms->name);
+
+ char req_buff[sizeof(union __RequestUnion__catch_mach_exc_subsystem) * 2];
+ char rep_buff[sizeof(union __ReplyUnion__catch_mach_exc_subsystem)];
+ mig_reply_error_t *req_hdr = (mig_reply_error_t *)&req_buff;
+ mig_reply_error_t *rep_hdr = (mig_reply_error_t *)&rep_buff;
+
+ mach_msg_return_t mr = ~MACH_MSG_SUCCESS;
+
+ do {
+ /* This should be a direct check on the Mach service to see if it's an exception-handling
+ * port, and it will break things if ReportCrash or SafetyNet start advertising other
+ * Mach services. But for now, it should be okay.
+ */
+ if (ms->job->alt_exc_handler || ms->job->internal_exc_handler) {
+ mr = launchd_exc_runtime_once(ms->port, sizeof(req_buff), sizeof(rep_buff), req_hdr, rep_hdr, 0);
+ } else {
+ mach_msg_options_t options = MACH_RCV_MSG |
+ MACH_RCV_TIMEOUT ;
+
+ mr = mach_msg((mach_msg_header_t *)req_hdr, options, 0, sizeof(req_buff), ms->port, 0, MACH_PORT_NULL);
+ switch (mr) {
+ case MACH_MSG_SUCCESS:
+ mach_msg_destroy((mach_msg_header_t *)req_hdr);
+ break;
+ case MACH_RCV_TIMED_OUT:
+ break;
+ case MACH_RCV_TOO_LARGE:
+ runtime_syslog(LOG_WARNING, "Tried to receive message that was larger than %lu bytes", sizeof(req_buff));
+ break;
+ default:
+ break;
+ }
+ }
+ } while (drain_all && mr != MACH_RCV_TIMED_OUT);
+}
+
+void
+machservice_delete(job_t j, struct machservice *ms, bool port_died)
+{
+ if (ms->alias) {
+ /* HACK: Egregious code duplication. But dealing with aliases is a
+ * pretty simple affair since they can't and shouldn't have any complex
+ * behaviors associated with them.
+ */
+ LIST_REMOVE(ms, name_hash_sle);
+ SLIST_REMOVE(&j->machservices, ms, machservice, sle);
+ free(ms);
+ return;
+ }
+
+ if (unlikely(ms->debug_on_close)) {
+ job_log(j, LOG_NOTICE, "About to enter kernel debugger because of Mach port: 0x%x", ms->port);
+ (void)job_assumes(j, host_reboot(mach_host_self(), HOST_REBOOT_DEBUGGER) == KERN_SUCCESS);
+ }
+
+ if (ms->recv && job_assumes(j, !machservice_active(ms))) {
+ job_log(j, LOG_DEBUG, "Closing receive right for %s", ms->name);
+ (void)job_assumes(j, launchd_mport_close_recv(ms->port) == KERN_SUCCESS);
+ }
+
+ (void)job_assumes(j, launchd_mport_deallocate(ms->port) == KERN_SUCCESS);
+
+ if (unlikely(ms->port == the_exception_server)) {
+ the_exception_server = 0;
+ }
+
+ job_log(j, LOG_DEBUG, "Mach service deleted%s: %s", port_died ? " (port died)" : "", ms->name);
+
+ if (ms->special_port_num) {
+ SLIST_REMOVE(&special_ports, ms, machservice, special_port_sle);
+ }
+ SLIST_REMOVE(&j->machservices, ms, machservice, sle);
+
+ if (!(j->dedicated_instance || ms->event_channel)) {
+ LIST_REMOVE(ms, name_hash_sle);
+ }
+ LIST_REMOVE(ms, port_hash_sle);
+
+ free(ms);
+}
+
+void
+machservice_request_notifications(struct machservice *ms)
+{
+ mach_msg_id_t which = MACH_NOTIFY_DEAD_NAME;
+
+ ms->isActive = true;
+
+ if (ms->recv) {
+ which = MACH_NOTIFY_PORT_DESTROYED;
+ job_checkin(ms->job);
+ }
+
+ (void)job_assumes(ms->job, launchd_mport_notify_req(ms->port, which) == KERN_SUCCESS);
+}
+
+#define NELEM(x) (sizeof(x)/sizeof(x[0]))
+#define END_OF(x) (&(x)[NELEM(x)])
+
+char **
+mach_cmd2argv(const char *string)
+{
+ char *argv[100], args[1000];
+ const char *cp;
+ char *argp = args, term, **argv_ret, *co;
+ unsigned int nargs = 0, i;
+
+ for (cp = string; *cp;) {
+ while (isspace(*cp))
+ cp++;
+ term = (*cp == '"') ? *cp++ : '\0';
+ if (nargs < NELEM(argv)) {
+ argv[nargs++] = argp;
+ }
+ while (*cp && (term ? *cp != term : !isspace(*cp)) && argp < END_OF(args)) {
+ if (*cp == '\\') {
+ cp++;
+ }
+ *argp++ = *cp;
+ if (*cp) {
+ cp++;
+ }
+ }
+ *argp++ = '\0';
+ }
+ argv[nargs] = NULL;
+
+ if (nargs == 0) {
+ return NULL;
+ }
+
+ argv_ret = malloc((nargs + 1) * sizeof(char *) + strlen(string) + 1);
+
+ if (!launchd_assumes(argv_ret != NULL)) {
+ return NULL;
+ }
+
+ co = (char *)argv_ret + (nargs + 1) * sizeof(char *);
+
+ for (i = 0; i < nargs; i++) {
+ strcpy(co, argv[i]);
+ argv_ret[i] = co;
+ co += strlen(argv[i]) + 1;
+ }
+ argv_ret[i] = NULL;
+
+ return argv_ret;
+}
+
+void
+job_checkin(job_t j)
+{
+ j->checkedin = true;
+}
+
+bool job_is_god(job_t j)
+{
+ return j->embedded_special_privileges;
+}
+
+bool
+job_ack_port_destruction(mach_port_t p)
+{
+ struct machservice *ms;
+ job_t j;
+
+ LIST_FOREACH(ms, &port_hash[HASH_PORT(p)], port_hash_sle) {
+ if (ms->recv && (ms->port == p)) {
+ break;
+ }
+ }
+
+ if (!jobmgr_assumes(root_jobmgr, ms != NULL)) {
+ return false;
+ }
+
+ j = ms->job;
+
+ jobmgr_log(root_jobmgr, LOG_DEBUG, "Receive right returned to us: %s", ms->name);
+
+ /* Without being the exception handler, NOTE_EXIT is our only way to tell if the job
+ * crashed, and we can't rely on NOTE_EXIT always being processed after all the job's
+ * receive rights have been returned.
+ *
+ * So when we get receive rights back, check to see if the job has been reaped yet. If
+ * not, then we add this service to a list of services to be drained on crash if it's
+ * requested that behavior. So, for a job with N receive rights all requesting that they
+ * be drained on crash, we can safely handle the following sequence of events.
+ *
+ * ReceiveRight0Returned
+ * ReceiveRight1Returned
+ * ReceiveRight2Returned
+ * NOTE_EXIT (reap, get exit status)
+ * ReceiveRight3Returned
+ * .
+ * .
+ * .
+ * ReceiveRight(N - 1)Returned
+ */
+
+ if (ms->drain_one_on_crash || ms->drain_all_on_crash) {
+ if (j->crashed && j->reaped) {
+ job_log(j, LOG_DEBUG, "Job has crashed. Draining port...");
+ machservice_drain_port(ms);
+ } else if (!(j->crashed || j->reaped)) {
+ job_log(j, LOG_DEBUG, "Job's exit status is still unknown. Deferring drain.");
+ }
+ }
+
+ /* If we get this notification after the job has been reaped, then we want to ping
+ * the event port to keep things going.
+ */
+ if (ms->event_update_port && !j->p && job_assumes(j, j->event_monitor)) {
+ if (_s_event_update_port == MACH_PORT_NULL) {
+ (void)job_assumes(j, launchd_mport_make_send_once(ms->port, &_s_event_update_port) == KERN_SUCCESS);
+ }
+ eventsystem_ping();
+ }
+
+ ms->isActive = false;
+ if (ms->delete_on_destruction) {
+ machservice_delete(j, ms, false);
+ } else if (ms->reset) {
+ machservice_resetport(j, ms);
+ }
+
+ job_dispatch(j, false);
+
+ root_jobmgr = jobmgr_do_garbage_collection(root_jobmgr);
+
+ return true;
+}
+
+void
+job_ack_no_senders(job_t j)
+{
+ j->priv_port_has_senders = false;
+
+ (void)job_assumes(j, launchd_mport_close_recv(j->j_port) == KERN_SUCCESS);
+ j->j_port = 0;
+
+ job_log(j, LOG_DEBUG, "No more senders on privileged Mach bootstrap port");
+
+ job_dispatch(j, false);
+}
+
+bool
+semaphoreitem_new(job_t j, semaphore_reason_t why, const char *what)
+{
+ struct semaphoreitem *si;
+ size_t alloc_sz = sizeof(struct semaphoreitem);
+
+ if (what) {
+ alloc_sz += strlen(what) + 1;
+ }
+
+ if (!job_assumes(j, si = calloc(1, alloc_sz))) {
+ return false;
+ }
+
+ si->fd = -1;
+ si->why = why;
+
+ if (what) {
+ strcpy(si->what_init, what);
+ }
+
+ SLIST_INSERT_HEAD(&j->semaphores, si, sle);
+
+ if ((why == OTHER_JOB_ENABLED || why == OTHER_JOB_DISABLED) && !j->nosy) {
+ job_log(j, LOG_DEBUG, "Job is interested in \"%s\".", what);
+ SLIST_INSERT_HEAD(&s_curious_jobs, j, curious_jobs_sle);
+ j->nosy = true;
+ }
+
+ semaphoreitem_runtime_mod_ref(si, true);
+
+ return true;
+}
+
+void
+semaphoreitem_runtime_mod_ref(struct semaphoreitem *si, bool add)
+{
+ /*
+ * External events need to be tracked.
+ * Internal events do NOT need to be tracked.
+ */
+
+ switch (si->why) {
+ case SUCCESSFUL_EXIT:
+ case FAILED_EXIT:
+ case OTHER_JOB_ENABLED:
+ case OTHER_JOB_DISABLED:
+ case OTHER_JOB_ACTIVE:
+ case OTHER_JOB_INACTIVE:
+ return;
+ default:
+ break;
+ }
+
+ if (add) {
+ runtime_add_weak_ref();
+ } else {
+ runtime_del_weak_ref();
+ }
+}
+
+void
+semaphoreitem_delete(job_t j, struct semaphoreitem *si)
+{
+ semaphoreitem_runtime_mod_ref(si, false);
+
+ SLIST_REMOVE(&j->semaphores, si, semaphoreitem, sle);
+
+ if (si->fd != -1) {
+ (void)job_assumes(j, runtime_close(si->fd) != -1);
+ }
+
+ /* We'll need to rethink this if it ever becomes possible to dynamically add or remove semaphores. */
+ if ((si->why == OTHER_JOB_ENABLED || si->why == OTHER_JOB_DISABLED) && j->nosy) {
+ j->nosy = false;
+ SLIST_REMOVE(&s_curious_jobs, j, job_s, curious_jobs_sle);
+ }
+
+ free(si);
+}
+
+void
+semaphoreitem_setup_dict_iter(launch_data_t obj, const char *key, void *context)
+{
+ struct semaphoreitem_dict_iter_context *sdic = context;
+ semaphore_reason_t why;
+
+ why = launch_data_get_bool(obj) ? sdic->why_true : sdic->why_false;
+
+ semaphoreitem_new(sdic->j, why, key);
+}
+
+void
+semaphoreitem_setup(launch_data_t obj, const char *key, void *context)
+{
+ struct semaphoreitem_dict_iter_context sdic = { context, 0, 0 };
+ job_t j = context;
+ semaphore_reason_t why;
+
+ switch (launch_data_get_type(obj)) {
+ case LAUNCH_DATA_BOOL:
+ if (strcasecmp(key, LAUNCH_JOBKEY_KEEPALIVE_NETWORKSTATE) == 0) {
+ why = launch_data_get_bool(obj) ? NETWORK_UP : NETWORK_DOWN;
+ semaphoreitem_new(j, why, NULL);
+ } else if (strcasecmp(key, LAUNCH_JOBKEY_KEEPALIVE_SUCCESSFULEXIT) == 0) {
+ why = launch_data_get_bool(obj) ? SUCCESSFUL_EXIT : FAILED_EXIT;
+ semaphoreitem_new(j, why, NULL);
+ j->start_pending = true;
+ } else if (strcasecmp(key, LAUNCH_JOBKEY_KEEPALIVE_AFTERINITIALDEMAND) == 0) {
+ j->needs_kickoff = launch_data_get_bool(obj);
+ } else if (strcasecmp(key, LAUNCH_JOBKEY_KEEPALIVE_CRASHED) == 0) {
+ why = launch_data_get_bool(obj) ? CRASHED : DID_NOT_CRASH;
+ semaphoreitem_new(j, why, NULL);
+ j->start_pending = true;
+ } else {
+ (void)job_assumes(j, false);
+ }
+ break;
+ case LAUNCH_DATA_DICTIONARY:
+ if (strcasecmp(key, LAUNCH_JOBKEY_KEEPALIVE_PATHSTATE) == 0) {
+ sdic.why_true = PATH_EXISTS;
+ sdic.why_false = PATH_MISSING;
+ } else if (strcasecmp(key, LAUNCH_JOBKEY_KEEPALIVE_OTHERJOBACTIVE) == 0) {
+ sdic.why_true = OTHER_JOB_ACTIVE;
+ sdic.why_false = OTHER_JOB_INACTIVE;
+ } else if (strcasecmp(key, LAUNCH_JOBKEY_KEEPALIVE_OTHERJOBENABLED) == 0) {
+ sdic.why_true = OTHER_JOB_ENABLED;
+ sdic.why_false = OTHER_JOB_DISABLED;
+ } else {
+ (void)job_assumes(j, false);
+ break;
+ }
+
+ launch_data_dict_iterate(obj, semaphoreitem_setup_dict_iter, &sdic);
+ break;
+ default:
+ (void)job_assumes(j, false);
+ break;
+ }
+}
+
+bool
+externalevent_new(job_t j, struct eventsystem *sys, char *evname, launch_data_t event)
+{
+ struct externalevent *ee = (struct externalevent *)calloc(1, sizeof(struct externalevent) + strlen(evname) + 1);
+ if (job_assumes(j, ee != NULL)) {
+ ee->event = launch_data_copy(event);
+ if (job_assumes(j, ee->event != NULL)) {
+ strcpy(ee->name, evname);
+ ee->job = j;
+ ee->id = sys->curid;
+ ee->sys = sys;
+ ee->state = false;
+ ee->wanted_state = true;
+ sys->curid++;
+
+ LIST_INSERT_HEAD(&j->events, ee, job_le);
+ LIST_INSERT_HEAD(&sys->events, ee, sys_le);
+
+ job_log(j, LOG_DEBUG, "New event: %s:%s", sys->name, evname);
+ } else {
+ free(ee);
+ ee = NULL;
+ }
+ }
+
+ eventsystem_ping();
+ return ee;
+}
+
+void
+externalevent_delete(struct externalevent *ee)
+{
+ launch_data_free(ee->event);
+ LIST_REMOVE(ee, job_le);
+ LIST_REMOVE(ee, sys_le);
+
+ free(ee);
+
+ eventsystem_ping();
+}
+
+void
+externalevent_setup(launch_data_t obj, const char *key, void *context)
+{
+ struct externalevent_iter_ctx *ctx = (struct externalevent_iter_ctx *)context;
+ (void)job_assumes(ctx->j, externalevent_new(ctx->j, ctx->sys, (char *)key, obj));
+}
+
+struct externalevent *
+externalevent_find(const char *sysname, uint64_t id)
+{
+ struct externalevent *ei = NULL;
+
+ struct eventsystem *es = eventsystem_find(sysname);
+ if (launchd_assumes(es != NULL)) {
+ LIST_FOREACH(ei, &es->events, sys_le) {
+ if (ei->id == id) {
+ break;
+ }
+ }
+ }
+
+ return ei;
+}
+
+struct eventsystem *
+eventsystem_new(const char *name)
+{
+ struct eventsystem *es = (struct eventsystem *)calloc(1, sizeof(struct eventsystem) + strlen(name) + 1);
+ if (launchd_assumes(es != NULL)) {
+ strcpy(es->name, name);
+ LIST_INSERT_HEAD(&_s_event_systems, es, global_le);
+ }
+
+ return es;
+}
+
+void
+eventsystem_delete(struct eventsystem *es)
+{
+ struct externalevent *ei = NULL;
+ while ((ei = LIST_FIRST(&es->events))) {
+ externalevent_delete(ei);
+ }
+
+ LIST_REMOVE(es, global_le);
+
+ free(es);
+}
+
+void
+eventsystem_setup(launch_data_t obj, const char *key, void *context)
+{
+ job_t j = (job_t)context;
+ if (!job_assumes(j, launch_data_get_type(obj) == LAUNCH_DATA_DICTIONARY)) {
+ return;
+ }
+
+ struct eventsystem *sys = eventsystem_find(key);
+ if (unlikely(sys == NULL)) {
+ sys = eventsystem_new(key);
+ job_log(j, LOG_DEBUG, "New event system: %s", key);
+ }
+
+ if (job_assumes(j, sys != NULL)) {
+ struct externalevent_iter_ctx ctx = {
+ .j = j,
+ .sys = sys,
+ };
+ launch_data_dict_iterate(obj, externalevent_setup, &ctx);
+ sys->has_updates = true;
+ }
+}
+
+struct eventsystem *
+eventsystem_find(const char *name)
+{
+ struct eventsystem *esi = NULL;
+ LIST_FOREACH(esi, &_s_event_systems, global_le) {
+ if (strcmp(name, esi->name) == 0) {
+ break;
+ }
+ }
+
+ return esi;
+}
+
+void
+eventsystem_ping(void)
+{
+ /* We don't wrap this in an assumes() macro because we could potentially
+ * call this function many times before the helper job gets back to us
+ * and gives us another send-once right. So if it's MACH_PORT_NULL, that
+ * means that we've sent a ping, but the helper hasn't yet checked in to
+ * get the new set of notifications.
+ */
+ if (_s_event_update_port != MACH_PORT_NULL) {
+ kern_return_t kr = helper_downcall_ping(_s_event_update_port);
+ if (kr != KERN_SUCCESS) {
+ runtime_syslog(LOG_NOTICE, "helper_downcall_ping(): kr = 0x%x", kr);
+ }
+ _s_event_update_port = MACH_PORT_NULL;
+ }
+}
+
+void
+jobmgr_dispatch_all_semaphores(jobmgr_t jm)
+{
+ jobmgr_t jmi, jmn;
+ job_t ji, jn;
+
+
+ SLIST_FOREACH_SAFE(jmi, &jm->submgrs, sle, jmn) {
+ jobmgr_dispatch_all_semaphores(jmi);
+ }
+
+ LIST_FOREACH_SAFE(ji, &jm->jobs, sle, jn) {
+ if (!SLIST_EMPTY(&ji->semaphores)) {
+ job_dispatch(ji, false);
+ }
+ }
+}
+
+time_t
+cronemu(int mon, int mday, int hour, int min)
+{
+ struct tm workingtm;
+ time_t now;
+
+ now = time(NULL);
+ workingtm = *localtime(&now);
+
+ workingtm.tm_isdst = -1;
+ workingtm.tm_sec = 0;
+ workingtm.tm_min++;
+
+ while (!cronemu_mon(&workingtm, mon, mday, hour, min)) {
+ workingtm.tm_year++;
+ workingtm.tm_mon = 0;
+ workingtm.tm_mday = 1;
+ workingtm.tm_hour = 0;
+ workingtm.tm_min = 0;
+ mktime(&workingtm);
+ }
+
+ return mktime(&workingtm);
+}
+
+time_t
+cronemu_wday(int wday, int hour, int min)
+{
+ struct tm workingtm;
+ time_t now;
+
+ now = time(NULL);
+ workingtm = *localtime(&now);
+
+ workingtm.tm_isdst = -1;
+ workingtm.tm_sec = 0;
+ workingtm.tm_min++;
+
+ if (wday == 7) {
+ wday = 0;
+ }
+
+ while (!(workingtm.tm_wday == wday && cronemu_hour(&workingtm, hour, min))) {
+ workingtm.tm_mday++;
+ workingtm.tm_hour = 0;
+ workingtm.tm_min = 0;
+ mktime(&workingtm);
+ }
+
+ return mktime(&workingtm);
+}
+
+bool
+cronemu_mon(struct tm *wtm, int mon, int mday, int hour, int min)
+{
+ if (mon == -1) {
+ struct tm workingtm = *wtm;
+ int carrytest;
+
+ while (!cronemu_mday(&workingtm, mday, hour, min)) {
+ workingtm.tm_mon++;
+ workingtm.tm_mday = 1;
+ workingtm.tm_hour = 0;
+ workingtm.tm_min = 0;
+ carrytest = workingtm.tm_mon;
+ mktime(&workingtm);
+ if (carrytest != workingtm.tm_mon) {
+ return false;
+ }
+ }
+ *wtm = workingtm;
+ return true;
+ }
+
+ if (mon < wtm->tm_mon) {
+ return false;
+ }
+
+ if (mon > wtm->tm_mon) {
+ wtm->tm_mon = mon;
+ wtm->tm_mday = 1;
+ wtm->tm_hour = 0;
+ wtm->tm_min = 0;
+ }
+
+ return cronemu_mday(wtm, mday, hour, min);
+}
+
+bool
+cronemu_mday(struct tm *wtm, int mday, int hour, int min)
+{
+ if (mday == -1) {
+ struct tm workingtm = *wtm;
+ int carrytest;
+
+ while (!cronemu_hour(&workingtm, hour, min)) {
+ workingtm.tm_mday++;
+ workingtm.tm_hour = 0;
+ workingtm.tm_min = 0;
+ carrytest = workingtm.tm_mday;
+ mktime(&workingtm);
+ if (carrytest != workingtm.tm_mday) {
+ return false;
+ }
+ }
+ *wtm = workingtm;
+ return true;
+ }
+
+ if (mday < wtm->tm_mday) {
+ return false;
+ }
+
+ if (mday > wtm->tm_mday) {
+ wtm->tm_mday = mday;
+ wtm->tm_hour = 0;
+ wtm->tm_min = 0;
+ }
+
+ return cronemu_hour(wtm, hour, min);
+}
+
+bool
+cronemu_hour(struct tm *wtm, int hour, int min)
+{
+ if (hour == -1) {
+ struct tm workingtm = *wtm;
+ int carrytest;
+
+ while (!cronemu_min(&workingtm, min)) {
+ workingtm.tm_hour++;
+ workingtm.tm_min = 0;
+ carrytest = workingtm.tm_hour;
+ mktime(&workingtm);
+ if (carrytest != workingtm.tm_hour) {
+ return false;
+ }
+ }
+ *wtm = workingtm;
+ return true;
+ }
+
+ if (hour < wtm->tm_hour) {
+ return false;
+ }
+
+ if (hour > wtm->tm_hour) {
+ wtm->tm_hour = hour;
+ wtm->tm_min = 0;
+ }
+
+ return cronemu_min(wtm, min);
+}
+
+bool
+cronemu_min(struct tm *wtm, int min)
+{
+ if (min == -1) {
+ return true;
+ }
+
+ if (min < wtm->tm_min) {
+ return false;
+ }
+
+ if (min > wtm->tm_min) {
+ wtm->tm_min = min;
+ }
+
+ return true;
+}
+
+kern_return_t
+job_mig_setup_shmem(job_t j, mach_port_t *shmem_port)
+{
+ memory_object_size_t size_of_page, size_of_page_orig;
+ vm_address_t vm_addr;
+ kern_return_t kr;
+
+ if (!launchd_assumes(j != NULL)) {
+ return BOOTSTRAP_NO_MEMORY;
+ }
+
+ if (unlikely(j->anonymous)) {
+ job_log(j, LOG_DEBUG, "Anonymous job tried to setup shared memory");
+ return BOOTSTRAP_NOT_PRIVILEGED;
+ }
+
+ if (unlikely(j->shmem)) {
+ job_log(j, LOG_ERR, "Tried to setup shared memory more than once");
+ return BOOTSTRAP_NOT_PRIVILEGED;
+ }
+
+ size_of_page_orig = size_of_page = getpagesize();
+
+ kr = vm_allocate(mach_task_self(), &vm_addr, size_of_page, true);
+
+ if (!job_assumes(j, kr == 0)) {
+ return kr;
+ }
+
+ j->shmem = (typeof(j->shmem))vm_addr;
+ j->shmem->vp_shmem_standby_timeout = j->timeout;
+
+ kr = mach_make_memory_entry_64(mach_task_self(), &size_of_page,
+ (memory_object_offset_t)vm_addr, VM_PROT_READ|VM_PROT_WRITE, shmem_port, 0);
+
+ if (job_assumes(j, kr == 0)) {
+ (void)job_assumes(j, size_of_page == size_of_page_orig);
+ }
+
+ /* no need to inherit this in child processes */
+ (void)job_assumes(j, vm_inherit(mach_task_self(), (vm_address_t)j->shmem, size_of_page_orig, VM_INHERIT_NONE) == 0);
+
+ return kr;
+}
+
+kern_return_t
+job_mig_create_server(job_t j, cmd_t server_cmd, uid_t server_uid, boolean_t on_demand, mach_port_t *server_portp)
+{
+ struct ldcred *ldc = runtime_get_caller_creds();
+ job_t js;
+
+ if (!launchd_assumes(j != NULL)) {
+ return BOOTSTRAP_NO_MEMORY;
+ }
+
+ if (unlikely(j->deny_job_creation)) {
+ return BOOTSTRAP_NOT_PRIVILEGED;
+ }
+
+#if HAVE_SANDBOX
+ const char **argv = (const char **)mach_cmd2argv(server_cmd);
+ if (unlikely(argv == NULL)) {
+ return BOOTSTRAP_NO_MEMORY;
+ }
+ if (unlikely(sandbox_check(ldc->pid, "job-creation", SANDBOX_FILTER_PATH, argv[0]) > 0)) {
+ free(argv);
+ return BOOTSTRAP_NOT_PRIVILEGED;
+ }
+ free(argv);
+#endif
+
+ job_log(j, LOG_DEBUG, "Server create attempt: %s", server_cmd);
+
+ if (pid1_magic) {
+ if (ldc->euid || ldc->uid) {
+ job_log(j, LOG_WARNING, "Server create attempt moved to per-user launchd: %s", server_cmd);
+ return VPROC_ERR_TRY_PER_USER;
+ }
+ } else {
+ if (unlikely(server_uid != getuid())) {
+ job_log(j, LOG_WARNING, "Server create: \"%s\": As UID %d, we will not be able to switch to UID %d",
+ server_cmd, getuid(), server_uid);
+ }
+ server_uid = 0; /* zero means "do nothing" */
+ }
+
+ js = job_new_via_mach_init(j, server_cmd, server_uid, on_demand);
+
+ if (unlikely(js == NULL)) {
+ return BOOTSTRAP_NO_MEMORY;
+ }
+
+ *server_portp = js->j_port;
+ return BOOTSTRAP_SUCCESS;
+}
+
+kern_return_t
+job_mig_send_signal(job_t j, mach_port_t srp, name_t targetlabel, int sig)
+{
+ struct ldcred *ldc = runtime_get_caller_creds();
+ job_t otherj;
+
+ if (!launchd_assumes(j != NULL)) {
+ return BOOTSTRAP_NO_MEMORY;
+ }
+
+ if (unlikely(ldc->euid != 0 && ldc->euid != getuid()) || j->deny_job_creation) {
+ #if TARGET_OS_EMBEDDED
+ if (!j->embedded_special_privileges) {
+ return BOOTSTRAP_NOT_PRIVILEGED;
+ }
+ #else
+ return BOOTSTRAP_NOT_PRIVILEGED;
+ #endif
+ }
+
+#if HAVE_SANDBOX
+ if (unlikely(sandbox_check(ldc->pid, "job-creation", SANDBOX_FILTER_NONE) > 0)) {
+ return BOOTSTRAP_NOT_PRIVILEGED;
+ }
+#endif
+
+ if (unlikely(!(otherj = job_find(NULL, targetlabel)))) {
+ return BOOTSTRAP_UNKNOWN_SERVICE;
+ }
+
+#if TARGET_OS_EMBEDDED
+ if (j->embedded_special_privileges && strcmp(j->username, otherj->username) != 0) {
+ return BOOTSTRAP_NOT_PRIVILEGED;
+ }
+#endif
+
+ if (sig == VPROC_MAGIC_UNLOAD_SIGNAL) {
+ bool do_block = otherj->p;
+
+ if (otherj->anonymous) {
+ return BOOTSTRAP_NOT_PRIVILEGED;
+ }
+
+ job_remove(otherj);
+
+ if (do_block) {
+ job_log(j, LOG_DEBUG, "Blocking MIG return of job_remove(): %s", otherj->label);
+ /* this is messy. We shouldn't access 'otherj' after job_remove(), but we check otherj->p first... */
+ (void)job_assumes(otherj, waiting4removal_new(otherj, srp));
+ return MIG_NO_REPLY;
+ } else {
+ return 0;
+ }
+ } else if (sig == VPROC_MAGIC_TRYKILL_SIGNAL) {
+ if (!j->kill_via_shmem) {
+ return BOOTSTRAP_NOT_PRIVILEGED;
+ }
+
+ if (!j->shmem) {
+ j->sent_kill_via_shmem = true;
+ (void)job_assumes(j, runtime_kill(otherj->p, SIGKILL) != -1);
+ return 0;
+ }
+
+#if !TARGET_OS_EMBEDDED
+ if (__sync_bool_compare_and_swap(&j->shmem->vp_shmem_transaction_cnt, 0, -1)) {
+ j->shmem->vp_shmem_flags |= VPROC_SHMEM_EXITING;
+ j->sent_kill_via_shmem = true;
+ (void)job_assumes(j, runtime_kill(otherj->p, SIGKILL) != -1);
+ return 0;
+ }
+#endif
+ return BOOTSTRAP_NOT_PRIVILEGED;
+ } else if (otherj->p) {
+ (void)job_assumes(j, runtime_kill(otherj->p, sig) != -1);
+ }
+
+ return 0;
+}
+
+kern_return_t
+job_mig_log_forward(job_t j, vm_offset_t inval, mach_msg_type_number_t invalCnt)
+{
+ struct ldcred *ldc = runtime_get_caller_creds();
+
+ if (!launchd_assumes(j != NULL)) {
+ return BOOTSTRAP_NO_MEMORY;
+ }
+
+ if (!job_assumes(j, j->per_user)) {
+ return BOOTSTRAP_NOT_PRIVILEGED;
+ }
+
+ return runtime_log_forward(ldc->euid, ldc->egid, inval, invalCnt);
+}
+
+kern_return_t
+job_mig_log_drain(job_t j, mach_port_t srp, vm_offset_t *outval, mach_msg_type_number_t *outvalCnt)
+{
+ struct ldcred *ldc = runtime_get_caller_creds();
+
+ if (!launchd_assumes(j != NULL)) {
+ return BOOTSTRAP_NO_MEMORY;
+ }
+
+ if (unlikely(ldc->euid)) {
+ return BOOTSTRAP_NOT_PRIVILEGED;
+ }
+
+ return runtime_log_drain(srp, outval, outvalCnt);
+}
+
+kern_return_t
+job_mig_swap_complex(job_t j, vproc_gsk_t inkey, vproc_gsk_t outkey, vm_offset_t inval, mach_msg_type_number_t invalCnt, vm_offset_t *outval, mach_msg_type_number_t *outvalCnt)
+{
+ const char *action;
+ launch_data_t input_obj = NULL, output_obj = NULL;
+ size_t data_offset = 0;
+ size_t packed_size;
+ struct ldcred *ldc = runtime_get_caller_creds();
+
+ if (!launchd_assumes(j != NULL)) {
+ return BOOTSTRAP_NO_MEMORY;
+ }
+ if (unlikely(inkey && ldc->euid && ldc->euid != getuid())) {
+ return BOOTSTRAP_NOT_PRIVILEGED;
+ }
+ if (unlikely(inkey && outkey && !job_assumes(j, inkey == outkey))) {
+ return 1;
+ }
+
+ if (inkey && outkey) {
+ action = "Swapping";
+ } else if (inkey) {
+ action = "Setting";
+ } else {
+ action = "Getting";
+ }
+
+ job_log(j, LOG_DEBUG, "%s key: %u", action, inkey ? inkey : outkey);
+
+ *outvalCnt = 20 * 1024 * 1024;
+ mig_allocate(outval, *outvalCnt);
+ if (!job_assumes(j, *outval != 0)) {
+ return 1;
+ }
+
+ /* Note to future maintainers: launch_data_unpack() does NOT return a heap object. The data
+ * is decoded in-place. So do not call launch_data_free() on input_obj.
+ */
+ runtime_ktrace0(RTKT_LAUNCHD_DATA_UNPACK);
+ if (unlikely(invalCnt && !job_assumes(j, (input_obj = launch_data_unpack((void *)inval, invalCnt, NULL, 0, &data_offset, NULL)) != NULL))) {
+ goto out_bad;
+ }
+
+ switch (outkey) {
+ case VPROC_GSK_ENVIRONMENT:
+ if (!job_assumes(j, (output_obj = launch_data_alloc(LAUNCH_DATA_DICTIONARY)))) {
+ goto out_bad;
+ }
+ jobmgr_export_env_from_other_jobs(j->mgr, output_obj);
+ runtime_ktrace0(RTKT_LAUNCHD_DATA_PACK);
+ if (!job_assumes(j, launch_data_pack(output_obj, (void *)*outval, *outvalCnt, NULL, NULL) != 0)) {
+ goto out_bad;
+ }
+ launch_data_free(output_obj);
+ break;
+ case VPROC_GSK_ALLJOBS:
+ if (!job_assumes(j, (output_obj = job_export_all()) != NULL)) {
+ goto out_bad;
+ }
+ ipc_revoke_fds(output_obj);
+ runtime_ktrace0(RTKT_LAUNCHD_DATA_PACK);
+ packed_size = launch_data_pack(output_obj, (void *)*outval, *outvalCnt, NULL, NULL);
+ if (!job_assumes(j, packed_size != 0)) {
+ goto out_bad;
+ }
+ launch_data_free(output_obj);
+ break;
+ case VPROC_GSK_MGR_NAME:
+ if (!job_assumes(j, (output_obj = launch_data_new_string(j->mgr->name)) != NULL)) {
+ goto out_bad;
+ }
+ packed_size = launch_data_pack(output_obj, (void *)*outval, *outvalCnt, NULL, NULL);
+ if (!job_assumes(j, packed_size != 0)) {
+ goto out_bad;
+ }
+
+ launch_data_free(output_obj);
+ break;
+ case VPROC_GSK_JOB_OVERRIDES_DB:
+ if (!job_assumes(j, (output_obj = launch_data_new_string(launchd_data_base_path(LAUNCHD_DB_TYPE_OVERRIDES))) != NULL)) {
+ goto out_bad;
+ }
+ packed_size = launch_data_pack(output_obj, (void *)*outval, *outvalCnt, NULL, NULL);
+ if (!job_assumes(j, packed_size != 0)) {
+ goto out_bad;
+ }
+
+ launch_data_free(output_obj);
+ break;
+ case VPROC_GSK_JOB_CACHE_DB:
+ if (!job_assumes(j, (output_obj = launch_data_new_string(launchd_data_base_path(LAUNCHD_DB_TYPE_JOBCACHE))) != NULL)) {
+ goto out_bad;
+ }
+ packed_size = launch_data_pack(output_obj, (void *)*outval, *outvalCnt, NULL, NULL);
+ if (!job_assumes(j, packed_size != 0)) {
+ goto out_bad;
+ }
+
+ job_log(j, LOG_DEBUG, "Location of job cache database: %s", launch_data_get_string(output_obj));
+
+ launch_data_free(output_obj);
+ break;
+ case 0:
+ mig_deallocate(*outval, *outvalCnt);
+ *outval = 0;
+ *outvalCnt = 0;
+ break;
+ default:
+ goto out_bad;
+ }
+
+ if (invalCnt) switch (inkey) {
+ case VPROC_GSK_ENVIRONMENT:
+ if (launch_data_get_type(input_obj) == LAUNCH_DATA_DICTIONARY) {
+ if (j->p) {
+ job_log(j, LOG_INFO, "Setting environment for a currently active job. This environment will take effect on the next invocation of the job.");
+ }
+ launch_data_dict_iterate(input_obj, envitem_setup_one_shot, j);
+ }
+ break;
+ case 0:
+ break;
+ default:
+ goto out_bad;
+ }
+
+ mig_deallocate(inval, invalCnt);
+ return 0;
+
+out_bad:
+ mig_deallocate(inval, invalCnt);
+ if (*outval) {
+ mig_deallocate(*outval, *outvalCnt);
+ }
+ if (output_obj) {
+ launch_data_free(output_obj);
+ }
+
+ return 1;
+}
+
+kern_return_t
+job_mig_swap_integer(job_t j, vproc_gsk_t inkey, vproc_gsk_t outkey, int64_t inval, int64_t *outval)
+{
+ const char *action;
+ kern_return_t kr = 0;
+ struct ldcred *ldc = runtime_get_caller_creds();
+ int oldmask;
+
+ if (!launchd_assumes(j != NULL)) {
+ return BOOTSTRAP_NO_MEMORY;
+ }
+
+ if (unlikely(inkey && ldc->euid && ldc->euid != getuid())) {
+ return BOOTSTRAP_NOT_PRIVILEGED;
+ }
+
+ if (unlikely(inkey && outkey && !job_assumes(j, inkey == outkey))) {
+ return 1;
+ }
+
+ if (inkey && outkey) {
+ action = "Swapping";
+ } else if (inkey) {
+ action = "Setting";
+ } else {
+ action = "Getting";
+ }
+
+ job_log(j, LOG_DEBUG, "%s key: %u", action, inkey ? inkey : outkey);
+
+ switch (outkey) {
+ case VPROC_GSK_ABANDON_PROCESS_GROUP:
+ *outval = j->abandon_pg;
+ break;
+ case VPROC_GSK_LAST_EXIT_STATUS:
+ *outval = j->last_exit_status;
+ break;
+ case VPROC_GSK_MGR_UID:
+ *outval = getuid();
+ break;
+ case VPROC_GSK_MGR_PID:
+ *outval = getpid();
+ break;
+ case VPROC_GSK_IS_MANAGED:
+ *outval = j->anonymous ? 0 : 1;
+ break;
+ case VPROC_GSK_BASIC_KEEPALIVE:
+ *outval = !j->ondemand;
+ break;
+ case VPROC_GSK_START_INTERVAL:
+ *outval = j->start_interval;
+ break;
+ case VPROC_GSK_IDLE_TIMEOUT:
+ *outval = j->timeout;
+ break;
+ case VPROC_GSK_EXIT_TIMEOUT:
+ *outval = j->exit_timeout;
+ break;
+ case VPROC_GSK_GLOBAL_LOG_MASK:
+ oldmask = runtime_setlogmask(LOG_UPTO(LOG_DEBUG));
+ *outval = oldmask;
+ runtime_setlogmask(oldmask);
+ break;
+ case VPROC_GSK_GLOBAL_UMASK:
+ oldmask = umask(0);
+ *outval = oldmask;
+ umask(oldmask);
+ break;
+ case VPROC_GSK_TRANSACTIONS_ENABLED:
+ job_log(j, LOG_DEBUG, "Reading transaction model status.");
+ *outval = j->kill_via_shmem;
+ break;
+ case VPROC_GSK_WAITFORDEBUGGER:
+ *outval = j->wait4debugger;
+ break;
+ case VPROC_GSK_EMBEDDEDROOTEQUIVALENT:
+ *outval = j->embedded_special_privileges;
+ break;
+ case 0:
+ *outval = 0;
+ break;
+ default:
+ kr = 1;
+ break;
+ }
+
+ switch (inkey) {
+ case VPROC_GSK_ABANDON_PROCESS_GROUP:
+ j->abandon_pg = (bool)inval;
+ break;
+ case VPROC_GSK_GLOBAL_ON_DEMAND:
+ job_log(j, LOG_DEBUG, "Job is setting global on-demand mode to %s (j->forced_peers_to_demand_mode = %s)", (bool)inval ? "true" : "false", j->forced_peers_to_demand_mode ? "true" : "false");
+ kr = job_set_global_on_demand(j, (bool)inval) ? 0 : 1;
+ break;
+ case VPROC_GSK_BASIC_KEEPALIVE:
+ j->ondemand = !inval;
+ break;
+ case VPROC_GSK_START_INTERVAL:
+ if (inval > UINT32_MAX || inval < 0) {
+ kr = 1;
+ } else if (inval) {
+ if (j->start_interval == 0) {
+ runtime_add_weak_ref();
+ }
+ j->start_interval = (typeof(j->start_interval)) inval;
+ (void)job_assumes(j, kevent_mod((uintptr_t)&j->start_interval, EVFILT_TIMER, EV_ADD, NOTE_SECONDS, j->start_interval, j) != -1);
+ } else if (j->start_interval) {
+ (void)job_assumes(j, kevent_mod((uintptr_t)&j->start_interval, EVFILT_TIMER, EV_DELETE, 0, 0, NULL) != -1);
+ if (j->start_interval != 0) {
+ runtime_del_weak_ref();
+ }
+ j->start_interval = 0;
+ }
+ break;
+ case VPROC_GSK_IDLE_TIMEOUT:
+ if (inval < 0 || inval > UINT32_MAX) {
+ kr = 1;
+ } else {
+ j->timeout = (typeof(j->timeout)) inval;
+ }
+ break;
+ case VPROC_GSK_EXIT_TIMEOUT:
+ if (inval < 0 || inval > UINT32_MAX) {
+ kr = 1;
+ } else {
+ j->exit_timeout = (typeof(j->exit_timeout)) inval;
+ }
+ break;
+ case VPROC_GSK_GLOBAL_LOG_MASK:
+ if (inval < 0 || inval > UINT32_MAX) {
+ kr = 1;
+ } else {
+ runtime_setlogmask((int) inval);
+ }
+ break;
+ case VPROC_GSK_GLOBAL_UMASK:
+ launchd_assert(sizeof (mode_t) == 2);
+ if (inval < 0 || inval > UINT16_MAX) {
+ kr = 1;
+ } else {
+#if HAVE_SANDBOX
+ if (unlikely(sandbox_check(ldc->pid, "job-creation", SANDBOX_FILTER_NONE) > 0)) {
+ kr = 1;
+ } else {
+ umask((mode_t) inval);
+ }
+#endif
+ }
+ break;
+ case VPROC_GSK_TRANSACTIONS_ENABLED:
+ if (!job_assumes(j, inval != 0)) {
+ job_log(j, LOG_WARNING, "Attempt to unregister from transaction model. This is not supported.");
+ kr = 1;
+ } else {
+ j->kill_via_shmem = (bool)inval;
+ }
+ break;
+ case VPROC_GSK_WEIRD_BOOTSTRAP:
+ if (job_assumes(j, j->weird_bootstrap)) {
+ job_log(j, LOG_DEBUG, "Unsetting weird bootstrap.");
+
+ mach_msg_size_t mxmsgsz = (typeof(mxmsgsz)) sizeof(union __RequestUnion__job_mig_protocol_vproc_subsystem);
+
+ if (job_mig_protocol_vproc_subsystem.maxsize > mxmsgsz) {
+ mxmsgsz = job_mig_protocol_vproc_subsystem.maxsize;
+ }
+
+ (void)job_assumes(j, runtime_add_mport(j->mgr->jm_port, protocol_vproc_server, mxmsgsz) == KERN_SUCCESS);
+ j->weird_bootstrap = false;
+ }
+ break;
+ case VPROC_GSK_WAITFORDEBUGGER:
+ j->wait4debugger_oneshot = inval;
+ break;
+ case VPROC_GSK_PERUSER_SUSPEND:
+ if (job_assumes(j, pid1_magic && ldc->euid == 0)) {
+ mach_port_t junk = MACH_PORT_NULL;
+ job_t jpu = jobmgr_lookup_per_user_context_internal(j, (uid_t)inval, &junk);
+ if (job_assumes(j, jpu != NULL)) {
+ struct suspended_peruser *spi = NULL;
+ LIST_FOREACH(spi, &j->suspended_perusers, sle) {
+ if ((int64_t)(spi->j->mach_uid) == inval) {
+ job_log(j, LOG_WARNING, "Job tried to suspend per-user launchd for UID %lli twice.", inval);
+ break;
+ }
+ }
+
+ if (spi == NULL) {
+ job_log(j, LOG_INFO, "Job is suspending the per-user launchd for UID %lli.", inval);
+ spi = (struct suspended_peruser *)calloc(sizeof(struct suspended_peruser), 1);
+ if (job_assumes(j, spi != NULL)) {
+ /* Stop listening for events.
+ *
+ * See <rdar://problem/9014146>.
+ */
+ if (jpu->peruser_suspend_count == 0) {
+ job_ignore(jpu);
+ }
+
+ spi->j = jpu;
+ spi->j->peruser_suspend_count++;
+ LIST_INSERT_HEAD(&j->suspended_perusers, spi, sle);
+ job_stop(spi->j);
+ *outval = jpu->p;
+ } else {
+ kr = BOOTSTRAP_NO_MEMORY;
+ }
+ }
+ }
+ } else {
+ kr = 1;
+ }
+ break;
+ case VPROC_GSK_PERUSER_RESUME:
+ if (job_assumes(j, pid1_magic == true)) {
+ struct suspended_peruser *spi = NULL, *spt = NULL;
+ LIST_FOREACH_SAFE(spi, &j->suspended_perusers, sle, spt) {
+ if ((int64_t)(spi->j->mach_uid) == inval) {
+ spi->j->peruser_suspend_count--;
+ LIST_REMOVE(spi, sle);
+ job_log(j, LOG_INFO, "Job is resuming the per-user launchd for UID %lli.", inval);
+ break;
+ }
+ }
+
+ if (!job_assumes(j, spi != NULL)) {
+ job_log(j, LOG_WARNING, "Job tried to resume per-user launchd for UID %lli that it did not suspend.", inval);
+ kr = BOOTSTRAP_NOT_PRIVILEGED;
+ } else if (spi->j->peruser_suspend_count == 0) {
+ job_watch(spi->j);
+ job_dispatch(spi->j, false);
+ free(spi);
+ }
+ } else {
+ kr = 1;
+ }
+ break;
+ case 0:
+ break;
+ default:
+ kr = 1;
+ break;
+ }
+
+ return kr;
+}
+
+kern_return_t
+job_mig_post_fork_ping(job_t j, task_t child_task, mach_port_t *asport)
+{
+ struct machservice *ms;
+
+ if (!launchd_assumes(j != NULL)) {
+ return BOOTSTRAP_NO_MEMORY;
+ }
+
+ job_log(j, LOG_DEBUG, "Post fork ping.");
+
+ job_setup_exception_port(j, child_task);
+
+ SLIST_FOREACH(ms, &special_ports, special_port_sle) {
+ if (j->per_user && (ms->special_port_num != TASK_ACCESS_PORT)) {
+ /* The TASK_ACCESS_PORT funny business is to workaround 5325399. */
+ continue;
+ }
+
+ errno = task_set_special_port(child_task, ms->special_port_num, ms->port);
+
+ if (unlikely(errno)) {
+ int desired_log_level = LOG_ERR;
+
+ if (j->anonymous) {
+ /* 5338127 */
+
+ desired_log_level = LOG_WARNING;
+
+ if (ms->special_port_num == TASK_SEATBELT_PORT) {
+ desired_log_level = LOG_DEBUG;
+ }
+ }
+
+ job_log(j, desired_log_level, "Could not setup Mach task special port %u: %s", ms->special_port_num, mach_error_string(errno));
+ }
+ }
+
+ /* MIG will not zero-initialize this pointer, so we must always do so. See
+ * <rdar://problem/8562593>.
+ */
+ *asport = MACH_PORT_NULL;
+#if !TARGET_OS_EMBEDDED
+ if (!j->anonymous) {
+ /* XPC services will spawn into the root security session by default.
+ * xpcproxy will switch them away if needed.
+ */
+ if (!(j->mgr->properties & BOOTSTRAP_PROPERTY_XPC_DOMAIN)) {
+ job_log(j, LOG_DEBUG, "Returning j->asport: %u", j->asport);
+ *asport = j->asport;
+ }
+ }
+#endif
+ (void)job_assumes(j, launchd_mport_deallocate(child_task) == KERN_SUCCESS);
+
+ return 0;
+}
+
+kern_return_t
+job_mig_reboot2(job_t j, uint64_t flags)
+{
+ char who_started_the_reboot[2048] = "";
+ struct proc_bsdshortinfo proc;
+ struct ldcred *ldc = runtime_get_caller_creds();
+ pid_t pid_to_log;
+
+ if (!launchd_assumes(j != NULL)) {
+ return BOOTSTRAP_NO_MEMORY;
+ }
+
+ if (unlikely(!pid1_magic)) {
+ return BOOTSTRAP_NOT_PRIVILEGED;
+ }
+
+#if !TARGET_OS_EMBEDDED
+ if (unlikely(ldc->euid)) {
+#else
+ if (unlikely(ldc->euid) && !j->embedded_special_privileges) {
+#endif
+ return BOOTSTRAP_NOT_PRIVILEGED;
+ }
+
+ for (pid_to_log = ldc->pid; pid_to_log; pid_to_log = proc.pbsi_ppid) {
+ size_t who_offset;
+ if (proc_pidinfo(pid_to_log, PROC_PIDT_SHORTBSDINFO, 1, &proc, PROC_PIDT_SHORTBSDINFO_SIZE) == 0) {
+ if (errno != ESRCH) {
+ job_assumes(j, errno == 0);
+ }
+ return 1;
+ }
+
+ if (!job_assumes(j, pid_to_log != (pid_t)proc.pbsi_ppid)) {
+ job_log(j, LOG_WARNING, "Job which is its own parent started reboot.");
+ snprintf(who_started_the_reboot, sizeof(who_started_the_reboot), "%s[%u]->%s[%u]->%s[%u]->...", proc.pbsi_comm, pid_to_log, proc.pbsi_comm, pid_to_log, proc.pbsi_comm, pid_to_log);
+ break;
+ }
+
+ who_offset = strlen(who_started_the_reboot);
+ snprintf(who_started_the_reboot + who_offset, sizeof(who_started_the_reboot) - who_offset,
+ " %s[%u]%s", proc.pbsi_comm, pid_to_log, proc.pbsi_ppid ? " ->" : "");
+ }
+
+ root_jobmgr->reboot_flags = (int)flags;
+ job_log(j, LOG_DEBUG, "reboot2() initiated by:%s", who_started_the_reboot);
+ launchd_shutdown();
+
+ return 0;
+}
+
+kern_return_t
+job_mig_getsocket(job_t j, name_t spr)
+{
+ if (!launchd_assumes(j != NULL)) {
+ return BOOTSTRAP_NO_MEMORY;
+ }
+
+ if (j->deny_job_creation) {
+ return BOOTSTRAP_NOT_PRIVILEGED;
+ }
+
+#if HAVE_SANDBOX
+ struct ldcred *ldc = runtime_get_caller_creds();
+ if (unlikely(sandbox_check(ldc->pid, "job-creation", SANDBOX_FILTER_NONE) > 0)) {
+ return BOOTSTRAP_NOT_PRIVILEGED;
+ }
+#endif
+
+ ipc_server_init();
+
+ if (unlikely(!sockpath)) {
+ return BOOTSTRAP_NO_MEMORY;
+ }
+
+ strncpy(spr, sockpath, sizeof(name_t));
+
+ return BOOTSTRAP_SUCCESS;
+}
+
+kern_return_t
+job_mig_log(job_t j, int pri, int err, logmsg_t msg)
+{
+ if (!launchd_assumes(j != NULL)) {
+ return BOOTSTRAP_NO_MEMORY;
+ }
+
+ if ((errno = err)) {
+ job_log_error(j, pri, "%s", msg);
+ } else {
+ job_log(j, pri, "%s", msg);
+ }
+
+ return 0;
+}
+
+job_t
+jobmgr_lookup_per_user_context_internal(job_t j, uid_t which_user, mach_port_t *mp)
+{
+ job_t ji = NULL;
+ LIST_FOREACH(ji, &root_jobmgr->jobs, sle) {
+ if (!ji->per_user) {
+ continue;
+ }
+ if (ji->mach_uid != which_user) {
+ continue;
+ }
+ if (SLIST_EMPTY(&ji->machservices)) {
+ continue;
+ }
+ if (!SLIST_FIRST(&ji->machservices)->per_user_hack) {
+ continue;
+ }
+ break;
+ }
+
+ if (unlikely(ji == NULL)) {
+ struct machservice *ms;
+ char lbuf[1024];
+
+ job_log(j, LOG_DEBUG, "Creating per user launchd job for UID: %u", which_user);
+
+ sprintf(lbuf, "com.apple.launchd.peruser.%u", which_user);
+
+ ji = job_new(root_jobmgr, lbuf, "/sbin/launchd", NULL);
+
+ if (ji != NULL) {
+ auditinfo_addr_t auinfo = {
+ .ai_termid = { .at_type = AU_IPv4 },
+ .ai_auid = which_user,
+ .ai_asid = AU_ASSIGN_ASID,
+ };
+
+ if (setaudit_addr(&auinfo, sizeof(auinfo)) == 0) {
+ job_log(ji, LOG_DEBUG, "Created new security session for per-user launchd: %u", auinfo.ai_asid);
+ (void)job_assumes(ji, (ji->asport = audit_session_self()) != MACH_PORT_NULL);
+
+ /* Kinda lame that we have to do this, but we can't create an
+ * audit session without joining it.
+ */
+ (void)job_assumes(ji, audit_session_join(g_audit_session_port));
+ ji->asid = auinfo.ai_asid;
+ } else {
+ job_log(ji, LOG_WARNING, "Could not set audit session!");
+ job_remove(ji);
+ return NULL;
+ }
+
+ ji->mach_uid = which_user;
+ ji->per_user = true;
+ ji->kill_via_shmem = true;
+
+ struct stat sb;
+ char pu_db[PATH_MAX];
+ snprintf(pu_db, sizeof(pu_db), LAUNCHD_DB_PREFIX "/%s", lbuf);
+
+ bool created = false;
+ int err = stat(pu_db, &sb);
+ if ((err == -1 && errno == ENOENT) || (err == 0 && !S_ISDIR(sb.st_mode))) {
+ if (err == 0) {
+ char move_aside[PATH_MAX];
+ snprintf(move_aside, sizeof(move_aside), LAUNCHD_DB_PREFIX "/%s.movedaside", lbuf);
+
+ (void)job_assumes(ji, rename(pu_db, move_aside) != -1);
+ }
+
+ (void)job_assumes(ji, mkdir(pu_db, S_IRWXU) != -1);
+ (void)job_assumes(ji, chown(pu_db, which_user, 0) != -1);
+ created = true;
+ }
+
+ if (!created) {
+ if (!job_assumes(ji, sb.st_uid == which_user)) {
+ (void)job_assumes(ji, chown(pu_db, which_user, 0) != -1);
+ }
+ if (!job_assumes(ji, sb.st_gid == 0)) {
+ (void)job_assumes(ji, chown(pu_db, which_user, 0) != -1);
+ }
+ if (!job_assumes(ji, sb.st_mode == (S_IRWXU | S_IFDIR))) {
+ (void)job_assumes(ji, chmod(pu_db, S_IRWXU) != -1);
+ }
+ }
+
+ if ((ms = machservice_new(ji, lbuf, mp, false)) == NULL) {
+ job_remove(ji);
+ ji = NULL;
+ } else {
+ ms->per_user_hack = true;
+ ms->hide = true;
+
+ ji = job_dispatch(ji, false);
+ }
+ }
+ } else {
+ *mp = machservice_port(SLIST_FIRST(&ji->machservices));
+ job_log(j, LOG_DEBUG, "Per user launchd job found for UID: %u", which_user);
+ }
+
+ return ji;
+}
+
+kern_return_t
+job_mig_lookup_per_user_context(job_t j, uid_t which_user, mach_port_t *up_cont)
+{
+ struct ldcred *ldc = runtime_get_caller_creds();
+ job_t jpu;
+
+#if TARGET_OS_EMBEDDED
+ /* There is no need for per-user launchd's on embedded. */
+ job_log(j, LOG_ERR, "Per-user launchds are not supported on this platform.");
+ return BOOTSTRAP_NOT_PRIVILEGED;
+#endif
+
+#if HAVE_SANDBOX
+ if (unlikely(sandbox_check(ldc->pid, "mach-per-user-lookup", SANDBOX_FILTER_NONE) > 0)) {
+ return BOOTSTRAP_NOT_PRIVILEGED;
+ }
+#endif
+
+ if (!launchd_assumes(j != NULL)) {
+ return BOOTSTRAP_NO_MEMORY;
+ }
+
+ job_log(j, LOG_INFO, "Looking up per user launchd for UID: %u", which_user);
+
+ if (unlikely(!pid1_magic)) {
+ job_log(j, LOG_ERR, "Only PID 1 supports per user launchd lookups.");
+ return BOOTSTRAP_NOT_PRIVILEGED;
+ }
+
+ if (ldc->euid || ldc->uid) {
+ which_user = ldc->euid ?: ldc->uid;
+ }
+
+ *up_cont = MACH_PORT_NULL;
+
+ jpu = jobmgr_lookup_per_user_context_internal(j, which_user, up_cont);
+
+ return 0;
+}
+
+kern_return_t
+job_mig_check_in2(job_t j, name_t servicename, mach_port_t *serviceportp, uuid_t instance_id, uint64_t flags)
+{
+ bool per_pid_service = flags & BOOTSTRAP_PER_PID_SERVICE;
+ bool strict = flags & BOOTSTRAP_STRICT_CHECKIN;
+ struct ldcred *ldc = runtime_get_caller_creds();
+ struct machservice *ms = NULL;
+ job_t jo;
+
+ if (!launchd_assumes(j != NULL)) {
+ return BOOTSTRAP_NO_MEMORY;
+ }
+
+ if (j->dedicated_instance) {
+ struct machservice *msi = NULL;
+ SLIST_FOREACH(msi, &j->machservices, sle) {
+ if (strncmp(servicename, msi->name, sizeof(name_t) - 1) == 0) {
+ uuid_copy(instance_id, j->instance_id);
+ ms = msi;
+ break;
+ }
+ }
+ } else {
+ ms = jobmgr_lookup_service(j->mgr, servicename, false, per_pid_service ? ldc->pid : 0);
+ }
+
+ if (strict) {
+ if (likely(ms != NULL)) {
+ if (ms->job != j) {
+ return BOOTSTRAP_NOT_PRIVILEGED;
+ } else if (ms->isActive) {
+ return BOOTSTRAP_SERVICE_ACTIVE;
+ }
+ } else {
+ return BOOTSTRAP_UNKNOWN_SERVICE;
+ }
+ } else if (ms == NULL) {
+ if (job_assumes(j, !j->dedicated_instance)) {
+ *serviceportp = MACH_PORT_NULL;
+
+ if (unlikely((ms = machservice_new(j, servicename, serviceportp, per_pid_service)) == NULL)) {
+ return BOOTSTRAP_NO_MEMORY;
+ }
+
+ /* Treat this like a legacy job. */
+ if (!j->legacy_mach_job) {
+ ms->isActive = true;
+ ms->recv = false;
+ }
+
+ if (!(j->anonymous || j->legacy_LS_job || j->legacy_mach_job)) {
+ job_log(j, LOG_SCOLDING, "Please add the following service to the configuration file for this job: %s", servicename);
+ }
+ } else {
+ return BOOTSTRAP_UNKNOWN_SERVICE;
+ }
+ } else {
+ if (unlikely((jo = machservice_job(ms)) != j)) {
+ static pid_t last_warned_pid;
+
+ if (last_warned_pid != ldc->pid) {
+ job_log(jo, LOG_WARNING, "The following job tried to hijack the service \"%s\" from this job: %s", servicename, j->label);
+ last_warned_pid = ldc->pid;
+ }
+
+ return BOOTSTRAP_NOT_PRIVILEGED;
+ }
+ if (unlikely(machservice_active(ms))) {
+ job_log(j, LOG_WARNING, "Check-in of Mach service failed. Already active: %s", servicename);
+ return BOOTSTRAP_SERVICE_ACTIVE;
+ }
+ }
+
+ job_checkin(j);
+ machservice_request_notifications(ms);
+
+ job_log(j, LOG_INFO, "Check-in of service: %s", servicename);
+
+ *serviceportp = machservice_port(ms);
+ return BOOTSTRAP_SUCCESS;
+}
+
+kern_return_t
+job_mig_register2(job_t j, name_t servicename, mach_port_t serviceport, uint64_t flags)
+{
+ struct machservice *ms;
+ struct ldcred *ldc = runtime_get_caller_creds();
+
+ if (!launchd_assumes(j != NULL)) {
+ return BOOTSTRAP_NO_MEMORY;
+ }
+
+ if (!(flags & BOOTSTRAP_PER_PID_SERVICE) && !j->legacy_LS_job) {
+ job_log(j, LOG_SCOLDING, "Performance: bootstrap_register() is deprecated. Service: %s", servicename);
+ }
+
+ job_log(j, LOG_DEBUG, "%sMach service registration attempt: %s", flags & BOOTSTRAP_PER_PID_SERVICE ? "Per PID " : "", servicename);
+
+ /* 5641783 for the embedded hack */
+#if !TARGET_OS_EMBEDDED
+ /*
+ * From a per-user/session launchd's perspective, SecurityAgent (UID
+ * 92) is a rogue application (not our UID, not root and not a child of
+ * us). We'll have to reconcile this design friction at a later date.
+ */
+ if (unlikely(j->anonymous && j->mgr->parentmgr == NULL && ldc->uid != 0 && ldc->uid != getuid() && ldc->uid != 92)) {
+ if (pid1_magic) {
+ return VPROC_ERR_TRY_PER_USER;
+ } else {
+ return BOOTSTRAP_NOT_PRIVILEGED;
+ }
+ }
+#endif
+
+ ms = jobmgr_lookup_service(j->mgr, servicename, false, flags & BOOTSTRAP_PER_PID_SERVICE ? ldc->pid : 0);
+
+ if (unlikely(ms)) {
+ if (machservice_job(ms) != j) {
+ return BOOTSTRAP_NOT_PRIVILEGED;
+ }
+ if (machservice_active(ms)) {
+ job_log(j, LOG_DEBUG, "Mach service registration failed. Already active: %s", servicename);
+ return BOOTSTRAP_SERVICE_ACTIVE;
+ }
+ if (ms->recv && (serviceport != MACH_PORT_NULL)) {
+ job_log(j, LOG_ERR, "bootstrap_register() erroneously called instead of bootstrap_check_in(). Mach service: %s", servicename);
+ return BOOTSTRAP_NOT_PRIVILEGED;
+ }
+ job_checkin(j);
+ machservice_delete(j, ms, false);
+ }
+
+ if (likely(serviceport != MACH_PORT_NULL)) {
+ if (likely(ms = machservice_new(j, servicename, &serviceport, flags & BOOTSTRAP_PER_PID_SERVICE ? true : false))) {
+ machservice_request_notifications(ms);
+ } else {
+ return BOOTSTRAP_NO_MEMORY;
+ }
+ }
+
+
+ return BOOTSTRAP_SUCCESS;
+}
+
+kern_return_t
+job_mig_look_up2(job_t j, mach_port_t srp, name_t servicename, mach_port_t *serviceportp, pid_t target_pid, uuid_t instance_id, uint64_t flags)
+{
+ struct machservice *ms = NULL;
+ struct ldcred *ldc = runtime_get_caller_creds();
+ kern_return_t kr;
+ bool per_pid_lookup = flags & BOOTSTRAP_PER_PID_SERVICE;
+ bool specific_instance = flags & BOOTSTRAP_SPECIFIC_INSTANCE;
+ bool strict_lookup = flags & BOOTSTRAP_STRICT_LOOKUP;
+ bool privileged = flags & BOOTSTRAP_PRIVILEGED_SERVER;
+
+ if (!launchd_assumes(j != NULL)) {
+ return BOOTSTRAP_NO_MEMORY;
+ }
+
+ bool xpc_req = j->mgr->properties & BOOTSTRAP_PROPERTY_XPC_DOMAIN;
+
+ /* 5641783 for the embedded hack */
+#if !TARGET_OS_EMBEDDED
+ if (unlikely(pid1_magic && j->anonymous && j->mgr->parentmgr == NULL && ldc->uid != 0 && ldc->euid != 0)) {
+ return VPROC_ERR_TRY_PER_USER;
+ }
+#endif
+
+#if HAVE_SANDBOX
+ /* We don't do sandbox checking for XPC domains because, by definition, all
+ * the services within your domain should be accessibly to you.
+ */
+ if (!xpc_req && unlikely(sandbox_check(ldc->pid, "mach-lookup", per_pid_lookup ? SANDBOX_FILTER_LOCAL_NAME : SANDBOX_FILTER_GLOBAL_NAME, servicename) > 0)) {
+ return BOOTSTRAP_NOT_PRIVILEGED;
+ }
+#endif
+
+ if (per_pid_lookup) {
+ ms = jobmgr_lookup_service(j->mgr, servicename, false, target_pid);
+ } else {
+ if (xpc_req) {
+ /* Requests from XPC domains stay local. */
+ ms = jobmgr_lookup_service(j->mgr, servicename, false, 0);
+ } else {
+ /* A strict lookup which is privileged won't even bother trying to
+ * find a service if we're not hosting the root Mach bootstrap.
+ */
+ if (strict_lookup && privileged) {
+ if (inherited_bootstrap_port == MACH_PORT_NULL) {
+ ms = jobmgr_lookup_service(j->mgr, servicename, true, 0);
+ }
+ } else {
+ ms = jobmgr_lookup_service(j->mgr, servicename, true, 0);
+ }
+ }
+ }
+
+ if (likely(ms)) {
+ ms = ms->alias ? ms->alias : ms;
+ if (unlikely(specific_instance && ms->job->multiple_instances)) {
+ job_t ji = NULL;
+ job_t instance = NULL;
+ LIST_FOREACH(ji, &ms->job->subjobs, subjob_sle) {
+ if (uuid_compare(instance_id, ji->instance_id) == 0) {
+ instance = ji;
+ break;
+ }
+ }
+
+ if (unlikely(instance == NULL)) {
+ job_log(ms->job, LOG_DEBUG, "Creating new instance of job based on lookup of service %s", ms->name);
+ instance = job_new_subjob(ms->job, instance_id);
+ if (job_assumes(j, instance != NULL)) {
+ /* Disable this support for now. We only support having
+ * multi-instance jobs within private XPC domains.
+ */
+#if 0
+ /* If the job is multi-instance, in a singleton XPC domain
+ * and the request is not coming from within that singleton
+ * domain, we need to alias the new job into the requesting
+ * domain.
+ */
+ if (!j->mgr->xpc_singleton && xpc_req) {
+ (void)job_assumes(instance, job_new_alias(j->mgr, instance));
+ }
+#endif
+ job_dispatch(instance, false);
+ }
+ }
+
+ ms = NULL;
+ if (job_assumes(j, instance != NULL)) {
+ struct machservice *msi = NULL;
+ SLIST_FOREACH(msi, &instance->machservices, sle) {
+ /* sizeof(servicename) will return the size of a pointer, even though it's
+ * an array type, because when passing arrays as parameters in C, they
+ * implicitly degrade to pointers.
+ */
+ if (strncmp(servicename, msi->name, sizeof(name_t) - 1) == 0) {
+ ms = msi;
+ break;
+ }
+ }
+ }
+ } else {
+ if (machservice_hidden(ms) && !machservice_active(ms)) {
+ ms = NULL;
+ } else if (unlikely(ms->per_user_hack)) {
+ ms = NULL;
+ }
+ }
+ }
+
+ if (likely(ms)) {
+ (void)job_assumes(j, machservice_port(ms) != MACH_PORT_NULL);
+ job_log(j, LOG_DEBUG, "%sMach service lookup: %s", per_pid_lookup ? "Per PID " : "", servicename);
+
+ if (unlikely(!per_pid_lookup && j->lastlookup == ms && j->lastlookup_gennum == ms->gen_num && !j->per_user)) {
+ /* we need to think more about the per_pid_lookup logic before logging about repeated lookups */
+ job_log(j, LOG_DEBUG, "Performance: Please fix the framework that talks to \"%s\" to cache the Mach port for service: %s", ms->job->label, servicename);
+ }
+
+ j->lastlookup = ms;
+ j->lastlookup_gennum = ms->gen_num;
+
+ *serviceportp = machservice_port(ms);
+
+ kr = BOOTSTRAP_SUCCESS;
+ } else if (strict_lookup && !privileged) {
+ /* Hack: We need to simulate XPC's desire not to establish a hierarchy. So if
+ * XPC is doing the lookup, and it's not a privileged lookup, we won't forward.
+ * But if it is a privileged lookup (that is, was looked up in XPC_DOMAIN_LOCAL_SYSTEM)
+ * then we must forward.
+ */
+ return BOOTSTRAP_UNKNOWN_SERVICE;
+ } else if (inherited_bootstrap_port != MACH_PORT_NULL) {
+ /* Requests from within an XPC domain don't get forwarded. */
+ job_log(j, LOG_DEBUG, "Mach service lookup forwarded: %s", servicename);
+ /* Clients potentially check the audit token of the reply to verify that the returned send right is trustworthy. */
+ (void)job_assumes(j, vproc_mig_look_up2_forward(inherited_bootstrap_port, srp, servicename, target_pid, instance_id, flags) == 0);
+ /* The previous routine moved the reply port, we're forced to return MIG_NO_REPLY now */
+ return MIG_NO_REPLY;
+ } else if (pid1_magic && j->anonymous && ldc->euid >= 500 && strcasecmp(j->mgr->name, VPROCMGR_SESSION_LOGINWINDOW) == 0) {
+ /*
+ * 5240036 Should start background session when a lookup of CCacheServer occurs
+ *
+ * This is a total hack. We sniff out loginwindow session, and attempt to guess what it is up to.
+ * If we find a EUID that isn't root, we force it over to the per-user context.
+ */
+ return VPROC_ERR_TRY_PER_USER;
+ } else {
+ job_log(j, LOG_DEBUG, "%sMach service lookup failed: %s", per_pid_lookup ? "Per PID " : "", servicename);
+ kr = BOOTSTRAP_UNKNOWN_SERVICE;
+ }
+
+ return kr;
+}
+
+kern_return_t
+job_mig_parent(job_t j, mach_port_t srp, mach_port_t *parentport)
+{
+ if (!launchd_assumes(j != NULL)) {
+ return BOOTSTRAP_NO_MEMORY;
+ }
+
+ job_log(j, LOG_DEBUG, "Requested parent bootstrap port");
+ jobmgr_t jm = j->mgr;
+
+ if (jobmgr_parent(jm)) {
+ *parentport = jobmgr_parent(jm)->jm_port;
+ } else if (MACH_PORT_NULL == inherited_bootstrap_port) {
+ *parentport = jm->jm_port;
+ } else {
+ (void)job_assumes(j, vproc_mig_parent_forward(inherited_bootstrap_port, srp) == 0);
+ /* The previous routine moved the reply port, we're forced to return MIG_NO_REPLY now */
+ return MIG_NO_REPLY;
+ }
+ return BOOTSTRAP_SUCCESS;
+}
+
+kern_return_t
+job_mig_get_root_bootstrap(job_t j, mach_port_t *rootbsp)
+{
+ if (inherited_bootstrap_port == MACH_PORT_NULL) {
+ *rootbsp = root_jobmgr->jm_port;
+ (void)job_assumes(j, launchd_mport_make_send(root_jobmgr->jm_port) == KERN_SUCCESS);
+ } else {
+ *rootbsp = inherited_bootstrap_port;
+ (void)job_assumes(j, launchd_mport_copy_send(inherited_bootstrap_port) == KERN_SUCCESS);
+ }
+
+ return BOOTSTRAP_SUCCESS;
+}
+
+kern_return_t
+job_mig_info(job_t j, name_array_t *servicenamesp, unsigned int *servicenames_cnt, name_array_t *servicejobsp, unsigned int *servicejobs_cnt, bootstrap_status_array_t *serviceactivesp, unsigned int *serviceactives_cnt, uint64_t flags)
+{
+ name_array_t service_names = NULL;
+ name_array_t service_jobs = NULL;
+ bootstrap_status_array_t service_actives = NULL;
+ unsigned int cnt = 0, cnt2 = 0;
+ jobmgr_t jm;
+
+ if (!launchd_assumes(j != NULL)) {
+ return BOOTSTRAP_NO_MEMORY;
+ }
+
+ if (g_flat_mach_namespace) {
+ if ((j->mgr->properties & BOOTSTRAP_PROPERTY_EXPLICITSUBSET) || (flags & BOOTSTRAP_FORCE_LOCAL)) {
+ jm = j->mgr;
+ } else {
+ jm = root_jobmgr;
+ }
+ } else {
+ jm = j->mgr;
+ }
+
+ unsigned int i = 0;
+ struct machservice *msi = NULL;
+ for (i = 0; i < MACHSERVICE_HASH_SIZE; i++) {
+ LIST_FOREACH(msi, &jm->ms_hash[i], name_hash_sle) {
+ cnt += !msi->per_pid ? 1 : 0;
+ }
+ }
+
+ if (cnt == 0) {
+ goto out;
+ }
+
+ mig_allocate((vm_address_t *)&service_names, cnt * sizeof(service_names[0]));
+ if (!job_assumes(j, service_names != NULL)) {
+ goto out_bad;
+ }
+
+ mig_allocate((vm_address_t *)&service_jobs, cnt * sizeof(service_jobs[0]));
+ if (!job_assumes(j, service_jobs != NULL)) {
+ goto out_bad;
+ }
+
+ mig_allocate((vm_address_t *)&service_actives, cnt * sizeof(service_actives[0]));
+ if (!job_assumes(j, service_actives != NULL)) {
+ goto out_bad;
+ }
+
+ for (i = 0; i < MACHSERVICE_HASH_SIZE; i++) {
+ LIST_FOREACH(msi, &jm->ms_hash[i], name_hash_sle) {
+ if (!msi->per_pid) {
+ strlcpy(service_names[cnt2], machservice_name(msi), sizeof(service_names[0]));
+ msi = msi->alias ? msi->alias : msi;
+ if (msi->job->mgr->shortdesc) {
+ strlcpy(service_jobs[cnt2], msi->job->mgr->shortdesc, sizeof(service_jobs[0]));
+ } else {
+ strlcpy(service_jobs[cnt2], msi->job->label, sizeof(service_jobs[0]));
+ }
+ service_actives[cnt2] = machservice_status(msi);
+ cnt2++;
+ }
+ }
+ }
+
+ (void)job_assumes(j, cnt == cnt2);
+
+out:
+ *servicenamesp = service_names;
+ *servicejobsp = service_jobs;
+ *serviceactivesp = service_actives;
+ *servicenames_cnt = *servicejobs_cnt = *serviceactives_cnt = cnt;
+
+ return BOOTSTRAP_SUCCESS;
+
+out_bad:
+ if (service_names) {
+ mig_deallocate((vm_address_t)service_names, cnt * sizeof(service_names[0]));
+ }
+ if (service_jobs) {
+ mig_deallocate((vm_address_t)service_jobs, cnt * sizeof(service_jobs[0]));
+ }
+ if (service_actives) {
+ mig_deallocate((vm_address_t)service_actives, cnt * sizeof(service_actives[0]));
+ }
+
+ return BOOTSTRAP_NO_MEMORY;
+}
+
+kern_return_t
+job_mig_lookup_children(job_t j, mach_port_array_t *child_ports, mach_msg_type_number_t *child_ports_cnt, name_array_t *child_names, mach_msg_type_number_t *child_names_cnt, bootstrap_property_array_t *child_properties,mach_msg_type_number_t *child_properties_cnt)
+{
+ kern_return_t kr = BOOTSTRAP_NO_MEMORY;
+ if (!launchd_assumes(j != NULL)) {
+ return BOOTSTRAP_NO_MEMORY;
+ }
+
+ struct ldcred *ldc = runtime_get_caller_creds();
+
+ /* Only allow root processes to look up children, even if we're in the per-user launchd.
+ * Otherwise, this could be used to cross sessions, which counts as a security vulnerability
+ * in a non-flat namespace.
+ */
+ if (ldc->euid != 0) {
+ job_log(j, LOG_WARNING, "Attempt to look up children of bootstrap by unprivileged job.");
+ return BOOTSTRAP_NOT_PRIVILEGED;
+ }
+
+ unsigned int cnt = 0;
+
+ jobmgr_t jmr = j->mgr;
+ jobmgr_t jmi = NULL;
+ SLIST_FOREACH(jmi, &jmr->submgrs, sle) {
+ cnt++;
+ }
+
+ /* Find our per-user launchds if we're PID 1. */
+ job_t ji = NULL;
+ if (pid1_magic) {
+ LIST_FOREACH(ji, &jmr->jobs, sle) {
+ cnt += ji->per_user ? 1 : 0;
}
- case LAUNCH_DATA_BOOL:
- b = launch_data_get_bool(obj);
- if (strcasecmp(key, LAUNCH_JOBKEY_MACH_RESETATCLOSE) == 0) {
- ms->reset = b;
- } else if (strcasecmp(key, LAUNCH_JOBKEY_MACH_HIDEUNTILCHECKIN) == 0) {
- ms->hide = b;
- } else if (strcasecmp(key, LAUNCH_JOBKEY_MACH_EXCEPTIONSERVER) == 0) {
- launchd_assumes(task_set_exception_ports(mts, EXC_MASK_ALL, ms->port,
- EXCEPTION_STATE_IDENTITY, f) == KERN_SUCCESS);
- } else if (strcasecmp(key, LAUNCH_JOBKEY_MACH_KUNCSERVER) == 0) {
- ms->kUNCServer = b;
- launchd_assumes(host_set_UNDServer(mhp, ms->port) == KERN_SUCCESS);
+ }
+
+ if (cnt == 0) {
+ return BOOTSTRAP_NO_CHILDREN;
+ }
+
+ mach_port_array_t _child_ports = NULL;
+ mig_allocate((vm_address_t *)&_child_ports, cnt * sizeof(_child_ports[0]));
+ if (!job_assumes(j, _child_ports != NULL)) {
+ kr = BOOTSTRAP_NO_MEMORY;
+ goto out_bad;
+ }
+
+ name_array_t _child_names = NULL;
+ mig_allocate((vm_address_t *)&_child_names, cnt * sizeof(_child_names[0]));
+ if (!job_assumes(j, _child_names != NULL)) {
+ kr = BOOTSTRAP_NO_MEMORY;
+ goto out_bad;
+ }
+
+ bootstrap_property_array_t _child_properties = NULL;
+ mig_allocate((vm_address_t *)&_child_properties, cnt * sizeof(_child_properties[0]));
+ if (!job_assumes(j, _child_properties != NULL)) {
+ kr = BOOTSTRAP_NO_MEMORY;
+ goto out_bad;
+ }
+
+ unsigned int cnt2 = 0;
+ SLIST_FOREACH(jmi, &jmr->submgrs, sle) {
+ if (jobmgr_assumes(jmi, launchd_mport_make_send(jmi->jm_port) == KERN_SUCCESS)) {
+ _child_ports[cnt2] = jmi->jm_port;
+ } else {
+ _child_ports[cnt2] = MACH_PORT_NULL;
}
- break;
- default:
- break;
+
+ strlcpy(_child_names[cnt2], jmi->name, sizeof(_child_names[0]));
+ _child_properties[cnt2] = jmi->properties;
+
+ cnt2++;
+ }
+
+ if (pid1_magic) LIST_FOREACH( ji, &jmr->jobs, sle) {
+ if (ji->per_user) {
+ if (job_assumes(ji, SLIST_FIRST(&ji->machservices)->per_user_hack == true)) {
+ mach_port_t port = machservice_port(SLIST_FIRST(&ji->machservices));
+
+ if (job_assumes(ji, launchd_mport_copy_send(port) == KERN_SUCCESS)) {
+ _child_ports[cnt2] = port;
+ } else {
+ _child_ports[cnt2] = MACH_PORT_NULL;
+ }
+ } else {
+ _child_ports[cnt2] = MACH_PORT_NULL;
+ }
+
+ strlcpy(_child_names[cnt2], ji->label, sizeof(_child_names[0]));
+ _child_properties[cnt2] |= BOOTSTRAP_PROPERTY_PERUSER;
+
+ cnt2++;
+ }
+ }
+
+ *child_names_cnt = cnt;
+ *child_ports_cnt = cnt;
+ *child_properties_cnt = cnt;
+
+ *child_names = _child_names;
+ *child_ports = _child_ports;
+ *child_properties = _child_properties;
+
+ unsigned int i = 0;
+ for (i = 0; i < cnt; i++) {
+ job_log(j, LOG_DEBUG, "child_names[%u] = %s", i, (char *)_child_names[i]);
+ }
+
+ return BOOTSTRAP_SUCCESS;
+out_bad:
+ if (_child_ports) {
+ mig_deallocate((vm_address_t)_child_ports, cnt * sizeof(_child_ports[0]));
+ }
+
+ if (_child_names) {
+ mig_deallocate((vm_address_t)_child_names, cnt * sizeof(_child_ports[0]));
+ }
+
+ if (_child_properties) {
+ mig_deallocate((vm_address_t)_child_properties, cnt * sizeof(_child_properties[0]));
+ }
+
+ return kr;
+}
+
+kern_return_t
+job_mig_transaction_count_for_pid(job_t j, pid_t p, int32_t *cnt, boolean_t *condemned)
+{
+ kern_return_t kr = KERN_FAILURE;
+ struct ldcred *ldc = runtime_get_caller_creds();
+ if ((ldc->euid != geteuid()) && (ldc->euid != 0)) {
+ return BOOTSTRAP_NOT_PRIVILEGED;
+ }
+
+ job_t j_for_pid = jobmgr_find_by_pid_deep(j->mgr, p, false);
+ if (j_for_pid) {
+ if (j_for_pid->kill_via_shmem) {
+ if (j_for_pid->shmem) {
+ *cnt = j_for_pid->shmem->vp_shmem_transaction_cnt;
+ *condemned = j_for_pid->shmem->vp_shmem_flags & VPROC_SHMEM_EXITING;
+ *cnt += *condemned ? 1 : 0;
+ } else {
+ *cnt = 0;
+ *condemned = false;
+ }
+
+ kr = BOOTSTRAP_SUCCESS;
+ } else {
+ kr = BOOTSTRAP_NO_MEMORY;
+ }
+ } else {
+ kr = BOOTSTRAP_UNKNOWN_SERVICE;
+ }
+
+ return kr;
+}
+
+kern_return_t
+job_mig_pid_is_managed(job_t j __attribute__((unused)), pid_t p, boolean_t *managed)
+{
+ struct ldcred *ldc = runtime_get_caller_creds();
+ if ((ldc->euid != geteuid()) && (ldc->euid != 0)) {
+ return BOOTSTRAP_NOT_PRIVILEGED;
+ }
+
+ /* This is so loginwindow doesn't try to quit GUI apps that have been launched
+ * directly by launchd as agents.
+ */
+ job_t j_for_pid = jobmgr_find_by_pid_deep(root_jobmgr, p, false);
+ if (j_for_pid && !j_for_pid->anonymous && !j_for_pid->legacy_LS_job) {
+ *managed = true;
+ }
+
+ return BOOTSTRAP_SUCCESS;
+}
+
+kern_return_t
+job_mig_port_for_label(job_t j __attribute__((unused)), name_t label, mach_port_t *mp)
+{
+ struct ldcred *ldc = runtime_get_caller_creds();
+ kern_return_t kr = BOOTSTRAP_NOT_PRIVILEGED;
+
+#if HAVE_SANDBOX
+ if (unlikely(sandbox_check(ldc->pid, "job-creation", SANDBOX_FILTER_NONE) > 0)) {
+ return BOOTSTRAP_NOT_PRIVILEGED;
+ }
+#endif
+
+ mach_port_t _mp = MACH_PORT_NULL;
+ if (!j->deny_job_creation && (ldc->euid == 0 || ldc->euid == geteuid())) {
+ job_t target_j = job_find(NULL, label);
+ if (jobmgr_assumes(root_jobmgr, target_j != NULL)) {
+ if (target_j->j_port == MACH_PORT_NULL) {
+ (void)job_assumes(target_j, job_setup_machport(target_j) == true);
+ }
+
+ _mp = target_j->j_port;
+ kr = _mp != MACH_PORT_NULL ? BOOTSTRAP_SUCCESS : BOOTSTRAP_NO_MEMORY;
+ } else {
+ kr = BOOTSTRAP_NO_MEMORY;
+ }
+ }
+
+ *mp = _mp;
+ return kr;
+}
+
+#if !TARGET_OS_EMBEDDED
+kern_return_t
+job_mig_set_security_session(job_t j, uuid_t uuid, mach_port_t asport)
+{
+ uuid_string_t uuid_str;
+ uuid_unparse(uuid, uuid_str);
+ job_log(j, LOG_DEBUG, "Setting session %u for UUID %s...", asport, uuid_str);
+
+ job_t ji = NULL, jt = NULL;
+ LIST_FOREACH_SAFE(ji, &s_needing_sessions, sle, jt) {
+ uuid_string_t uuid_str2;
+ uuid_unparse(ji->expected_audit_uuid, uuid_str2);
+
+ if (uuid_compare(uuid, ji->expected_audit_uuid) == 0) {
+ uuid_clear(ji->expected_audit_uuid);
+ if (asport != MACH_PORT_NULL ) {
+ job_log(ji, LOG_DEBUG, "Job should join session with port %u", asport);
+ (void)job_assumes(j, launchd_mport_copy_send(asport) == KERN_SUCCESS);
+ } else {
+ job_log(ji, LOG_DEBUG, "No session to set for job. Using our session.");
+ }
+
+ ji->asport = asport;
+ LIST_REMOVE(ji, needing_session_sle);
+ job_dispatch(ji, false);
+ }
+ }
+
+ /* Each job that the session port was set for holds a reference. At the end of
+ * the loop, there will be one extra reference belonging to this MiG protocol.
+ * We need to release it so that the session goes away when all the jobs
+ * referencing it are unloaded.
+ */
+ (void)job_assumes(j, launchd_mport_deallocate(asport) == KERN_SUCCESS);
+
+ return KERN_SUCCESS;
+}
+#else
+kern_return_t
+job_mig_set_security_session(job_t j __attribute__((unused)), uuid_t uuid __attribute__((unused)), mach_port_t session __attribute__((unused)))
+{
+ return KERN_SUCCESS;
+}
+#endif
+
+jobmgr_t
+jobmgr_find_by_name(jobmgr_t jm, const char *where)
+{
+ jobmgr_t jmi, jmi2;
+
+ /* NULL is only passed for our custom API for LaunchServices. If that is the case, we do magic. */
+ if (where == NULL) {
+ if (strcasecmp(jm->name, VPROCMGR_SESSION_LOGINWINDOW) == 0) {
+ where = VPROCMGR_SESSION_LOGINWINDOW;
+ } else {
+ where = VPROCMGR_SESSION_AQUA;
+ }
+ }
+
+ if (strcasecmp(jm->name, where) == 0) {
+ return jm;
+ }
+
+ if (strcasecmp(where, VPROCMGR_SESSION_BACKGROUND) == 0 && !pid1_magic) {
+ jmi = root_jobmgr;
+ goto jm_found;
+ }
+
+ SLIST_FOREACH(jmi, &root_jobmgr->submgrs, sle) {
+ if (unlikely(jmi->shutting_down)) {
+ continue;
+ } else if (jmi->properties & BOOTSTRAP_PROPERTY_XPC_DOMAIN) {
+ continue;
+ } else if (strcasecmp(jmi->name, where) == 0) {
+ goto jm_found;
+ } else if (strcasecmp(jmi->name, VPROCMGR_SESSION_BACKGROUND) == 0 && pid1_magic) {
+ SLIST_FOREACH(jmi2, &jmi->submgrs, sle) {
+ if (strcasecmp(jmi2->name, where) == 0) {
+ jmi = jmi2;
+ goto jm_found;
+ }
+ }
+ }
+ }
+
+jm_found:
+ return jmi;
+}
+
+kern_return_t
+job_mig_move_subset(job_t j, mach_port_t target_subset, name_t session_type, mach_port_t asport, uint64_t flags)
+{
+ mach_msg_type_number_t l2l_i, l2l_port_cnt = 0;
+ mach_port_array_t l2l_ports = NULL;
+ mach_port_t reqport, rcvright;
+ kern_return_t kr = 1;
+ launch_data_t out_obj_array = NULL;
+ struct ldcred *ldc = runtime_get_caller_creds();
+ jobmgr_t jmr = NULL;
+
+ if (!launchd_assumes(j != NULL)) {
+ return BOOTSTRAP_NO_MEMORY;
+ }
+
+ if (job_mig_intran2(root_jobmgr, target_subset, ldc->pid)) {
+ job_log(j, LOG_ERR, "Moving a session to ourself is bogus.");
+
+ kr = BOOTSTRAP_NOT_PRIVILEGED;
+ goto out;
+ }
+
+ job_log(j, LOG_DEBUG, "Move subset attempt: 0x%x", target_subset);
+
+ kr = _vproc_grab_subset(target_subset, &reqport, &rcvright, &out_obj_array, &l2l_ports, &l2l_port_cnt);
+
+ if (!job_assumes(j, kr == 0)) {
+ goto out;
+ }
+
+ launchd_assert(launch_data_array_get_count(out_obj_array) == l2l_port_cnt);
+
+ if (!job_assumes(j, (jmr = jobmgr_new(j->mgr, reqport, rcvright, false, session_type, false, asport)) != NULL)) {
+ kr = BOOTSTRAP_NO_MEMORY;
+ goto out;
+ }
+
+ jmr->properties |= BOOTSTRAP_PROPERTY_MOVEDSUBSET;
+
+ /* This is a hack. We should be doing this in jobmgr_new(), but since we're in the middle of
+ * processing an IPC request, we'll do this action before the new job manager can get any IPC
+ * requests. This serialization is guaranteed since we are single-threaded in that respect.
+ */
+ if (flags & LAUNCH_GLOBAL_ON_DEMAND) {
+ /* This is so awful. */
+ /* Remove the job from its current job manager. */
+ LIST_REMOVE(j, sle);
+ LIST_REMOVE(j, pid_hash_sle);
+
+ /* Put the job into the target job manager. */
+ LIST_INSERT_HEAD(&jmr->jobs, j, sle);
+ LIST_INSERT_HEAD(&jmr->active_jobs[ACTIVE_JOB_HASH(j->p)], j, pid_hash_sle);
+
+ j->mgr = jmr;
+ job_set_global_on_demand(j, true);
+
+ if (!j->holds_ref) {
+ j->holds_ref = true;
+ runtime_add_ref();
+ }
+ }
+
+ for (l2l_i = 0; l2l_i < l2l_port_cnt; l2l_i++) {
+ launch_data_t tmp, obj_at_idx;
+ struct machservice *ms;
+ job_t j_for_service;
+ const char *serv_name;
+ pid_t target_pid;
+ bool serv_perpid;
+
+ (void)job_assumes(j, obj_at_idx = launch_data_array_get_index(out_obj_array, l2l_i));
+ (void)job_assumes(j, tmp = launch_data_dict_lookup(obj_at_idx, TAKE_SUBSET_PID));
+ target_pid = (pid_t)launch_data_get_integer(tmp);
+ (void)job_assumes(j, tmp = launch_data_dict_lookup(obj_at_idx, TAKE_SUBSET_PERPID));
+ serv_perpid = launch_data_get_bool(tmp);
+ (void)job_assumes(j, tmp = launch_data_dict_lookup(obj_at_idx, TAKE_SUBSET_NAME));
+ serv_name = launch_data_get_string(tmp);
+
+ j_for_service = jobmgr_find_by_pid(jmr, target_pid, true);
+
+ if (unlikely(!j_for_service)) {
+ /* The PID probably exited */
+ (void)job_assumes(j, launchd_mport_deallocate(l2l_ports[l2l_i]) == KERN_SUCCESS);
+ continue;
+ }
+
+ if (likely(ms = machservice_new(j_for_service, serv_name, &l2l_ports[l2l_i], serv_perpid))) {
+ job_log(j, LOG_DEBUG, "Importing %s into new bootstrap.", serv_name);
+ machservice_request_notifications(ms);
+ }
+ }
+
+ kr = 0;
+
+out:
+ if (out_obj_array) {
+ launch_data_free(out_obj_array);
+ }
+
+ if (l2l_ports) {
+ mig_deallocate((vm_address_t)l2l_ports, l2l_port_cnt * sizeof(l2l_ports[0]));
+ }
+
+ if (kr == 0) {
+ if (target_subset) {
+ (void)job_assumes(j, launchd_mport_deallocate(target_subset) == KERN_SUCCESS);
+ }
+ if (asport) {
+ (void)job_assumes(j, launchd_mport_deallocate(asport) == KERN_SUCCESS);
+ }
+ } else if (jmr) {
+ jobmgr_shutdown(jmr);
+ }
+
+ return kr;
+}
+
+kern_return_t
+job_mig_init_session(job_t j, name_t session_type, mach_port_t asport)
+{
+ job_t j2;
+
+ kern_return_t kr = BOOTSTRAP_NO_MEMORY;
+ if (j->mgr->session_initialized) {
+ job_log(j, LOG_ERR, "Tried to initialize an already setup session!");
+ kr = BOOTSTRAP_NOT_PRIVILEGED;
+ } else if (strcmp(session_type, VPROCMGR_SESSION_LOGINWINDOW) == 0) {
+ jobmgr_t jmi;
+
+ /*
+ * 5330262
+ *
+ * We're working around LoginWindow and the WindowServer.
+ *
+ * In practice, there is only one LoginWindow session. Unfortunately, for certain
+ * scenarios, the WindowServer spawns loginwindow, and in those cases, it frequently
+ * spawns a replacement loginwindow session before cleaning up the previous one.
+ *
+ * We're going to use the creation of a new LoginWindow context as a clue that the
+ * previous LoginWindow context is on the way out and therefore we should just
+ * kick-start the shutdown of it.
+ */
+
+ SLIST_FOREACH(jmi, &root_jobmgr->submgrs, sle) {
+ if (unlikely(jmi->shutting_down)) {
+ continue;
+ } else if (strcasecmp(jmi->name, session_type) == 0) {
+ jobmgr_shutdown(jmi);
+ break;
+ }
+ }
+ }
+
+ jobmgr_log(j->mgr, LOG_DEBUG, "Initializing as %s", session_type);
+ strcpy(j->mgr->name_init, session_type);
+
+ if (job_assumes(j, (j2 = jobmgr_init_session(j->mgr, session_type, false)))) {
+ j2->asport = asport;
+ (void)job_assumes(j, job_dispatch(j2, true));
+ kr = BOOTSTRAP_SUCCESS;
+ }
+
+ return kr;
+}
+
+kern_return_t
+job_mig_switch_to_session(job_t j, mach_port_t requestor_port, name_t session_name, mach_port_t asport, mach_port_t *new_bsport)
+{
+ struct ldcred *ldc = runtime_get_caller_creds();
+ if (!jobmgr_assumes(root_jobmgr, j != NULL)) {
+ jobmgr_log(root_jobmgr, LOG_ERR, "%s() called with NULL job: PID %d", __func__, ldc->pid);
+ return BOOTSTRAP_NO_MEMORY;
+ }
+
+ job_log(j, LOG_DEBUG, "Job wants to move to %s session.", session_name);
+
+ if (!job_assumes(j, pid1_magic == false)) {
+ job_log(j, LOG_WARNING, "Switching sessions is not allowed in the system Mach bootstrap.");
+ return BOOTSTRAP_NOT_PRIVILEGED;
+ }
+
+ if (!j->anonymous) {
+ job_log(j, LOG_NOTICE, "Non-anonymous job tried to switch sessions. Please use LimitLoadToSessionType instead.");
+ return BOOTSTRAP_NOT_PRIVILEGED;
+ }
+
+ jobmgr_t target_jm = jobmgr_find_by_name(root_jobmgr, session_name);
+ if (target_jm == j->mgr) {
+ job_log(j, LOG_DEBUG, "Job is already in its desired session (%s).", session_name);
+ *new_bsport = target_jm->jm_port;
+ return BOOTSTRAP_SUCCESS;
+ }
+
+ if (!target_jm) {
+ target_jm = jobmgr_new(j->mgr, requestor_port, MACH_PORT_NULL, false, session_name, false, asport);
+ if (target_jm) {
+ target_jm->properties |= BOOTSTRAP_PROPERTY_IMPLICITSUBSET;
+ (void)job_assumes(j, launchd_mport_deallocate(asport) == KERN_SUCCESS);
+ }
+ }
+
+ if (!job_assumes(j, target_jm != NULL)) {
+ job_log(j, LOG_WARNING, "Could not find %s session!", session_name);
+ return BOOTSTRAP_NO_MEMORY;
+ }
+
+ /* Remove the job from it's current job manager. */
+ LIST_REMOVE(j, sle);
+ LIST_REMOVE(j, pid_hash_sle);
+
+ job_t ji = NULL, jit = NULL;
+ LIST_FOREACH_SAFE(ji, &j->mgr->global_env_jobs, global_env_sle, jit) {
+ if (ji == j) {
+ LIST_REMOVE(ji, global_env_sle);
+ break;
+ }
+ }
+
+ /* Put the job into the target job manager. */
+ LIST_INSERT_HEAD(&target_jm->jobs, j, sle);
+ LIST_INSERT_HEAD(&target_jm->active_jobs[ACTIVE_JOB_HASH(j->p)], j, pid_hash_sle);
+
+ if (ji) {
+ LIST_INSERT_HEAD(&target_jm->global_env_jobs, j, global_env_sle);
+ }
+
+ /* Move our Mach services over if we're not in a flat namespace. */
+ if (!g_flat_mach_namespace && !SLIST_EMPTY(&j->machservices)) {
+ struct machservice *msi = NULL, *msit = NULL;
+ SLIST_FOREACH_SAFE(msi, &j->machservices, sle, msit) {
+ LIST_REMOVE(msi, name_hash_sle);
+ LIST_INSERT_HEAD(&target_jm->ms_hash[hash_ms(msi->name)], msi, name_hash_sle);
+ }
+ }
+
+ j->mgr = target_jm;
+
+ if (!j->holds_ref) {
+ /* Anonymous jobs which move around are particularly interesting to us, so we want to
+ * stick around while they're still around.
+ * For example, login calls into the PAM launchd module, which moves the process into
+ * the StandardIO session by default. So we'll hold a reference on that job to prevent
+ * ourselves from going away.
+ */
+ j->holds_ref = true;
+ runtime_add_ref();
+ }
+
+ *new_bsport = target_jm->jm_port;
+
+ return KERN_SUCCESS;
+}
+
+kern_return_t
+job_mig_take_subset(job_t j, mach_port_t *reqport, mach_port_t *rcvright,
+ vm_offset_t *outdata, mach_msg_type_number_t *outdataCnt,
+ mach_port_array_t *portsp, unsigned int *ports_cnt)
+{
+ launch_data_t tmp_obj, tmp_dict, outdata_obj_array = NULL;
+ mach_port_array_t ports = NULL;
+ unsigned int cnt = 0, cnt2 = 0;
+ size_t packed_size;
+ struct machservice *ms;
+ jobmgr_t jm;
+ job_t ji;
+
+ if (!launchd_assumes(j != NULL)) {
+ return BOOTSTRAP_NO_MEMORY;
+ }
+
+ jm = j->mgr;
+
+ if (unlikely(!pid1_magic)) {
+ job_log(j, LOG_ERR, "Only the system launchd will transfer Mach sub-bootstraps.");
+ return BOOTSTRAP_NOT_PRIVILEGED;
+ }
+ if (unlikely(jobmgr_parent(jm) == NULL)) {
+ job_log(j, LOG_ERR, "Root Mach bootstrap cannot be transferred.");
+ return BOOTSTRAP_NOT_PRIVILEGED;
+ }
+ if (unlikely(strcasecmp(jm->name, VPROCMGR_SESSION_AQUA) == 0)) {
+ job_log(j, LOG_ERR, "Cannot transfer a setup GUI session.");
+ return BOOTSTRAP_NOT_PRIVILEGED;
+ }
+ if (unlikely(!j->anonymous)) {
+ job_log(j, LOG_ERR, "Only the anonymous job can transfer Mach sub-bootstraps.");
+ return BOOTSTRAP_NOT_PRIVILEGED;
+ }
+
+ job_log(j, LOG_DEBUG, "Transferring sub-bootstrap to the per session launchd.");
+
+ outdata_obj_array = launch_data_alloc(LAUNCH_DATA_ARRAY);
+ if (!job_assumes(j, outdata_obj_array)) {
+ goto out_bad;
+ }
+
+ *outdataCnt = 20 * 1024 * 1024;
+ mig_allocate(outdata, *outdataCnt);
+ if (!job_assumes(j, *outdata != 0)) {
+ return 1;
+ }
+
+ LIST_FOREACH(ji, &j->mgr->jobs, sle) {
+ if (!ji->anonymous) {
+ continue;
+ }
+ SLIST_FOREACH(ms, &ji->machservices, sle) {
+ cnt++;
+ }
+ }
+
+ mig_allocate((vm_address_t *)&ports, cnt * sizeof(ports[0]));
+ if (!job_assumes(j, ports != NULL)) {
+ goto out_bad;
+ }
+
+ LIST_FOREACH(ji, &j->mgr->jobs, sle) {
+ if (!ji->anonymous) {
+ continue;
+ }
+
+ SLIST_FOREACH(ms, &ji->machservices, sle) {
+ if (job_assumes(j, (tmp_dict = launch_data_alloc(LAUNCH_DATA_DICTIONARY)))) {
+ (void)job_assumes(j, launch_data_array_set_index(outdata_obj_array, tmp_dict, cnt2));
+ } else {
+ goto out_bad;
+ }
+
+ if (job_assumes(j, (tmp_obj = launch_data_new_string(machservice_name(ms))))) {
+ (void)job_assumes(j, launch_data_dict_insert(tmp_dict, tmp_obj, TAKE_SUBSET_NAME));
+ } else {
+ goto out_bad;
+ }
+
+ if (job_assumes(j, (tmp_obj = launch_data_new_integer((ms->job->p))))) {
+ (void)job_assumes(j, launch_data_dict_insert(tmp_dict, tmp_obj, TAKE_SUBSET_PID));
+ } else {
+ goto out_bad;
+ }
+
+ if (job_assumes(j, (tmp_obj = launch_data_new_bool((ms->per_pid))))) {
+ (void)job_assumes(j, launch_data_dict_insert(tmp_dict, tmp_obj, TAKE_SUBSET_PERPID));
+ } else {
+ goto out_bad;
+ }
+
+ ports[cnt2] = machservice_port(ms);
+
+ /* Increment the send right by one so we can shutdown the jobmgr cleanly */
+ (void)jobmgr_assumes(jm, (errno = launchd_mport_copy_send(ports[cnt2])) == KERN_SUCCESS);
+ cnt2++;
+ }
+ }
+
+ (void)job_assumes(j, cnt == cnt2);
+
+ runtime_ktrace0(RTKT_LAUNCHD_DATA_PACK);
+ packed_size = launch_data_pack(outdata_obj_array, (void *)*outdata, *outdataCnt, NULL, NULL);
+ if (!job_assumes(j, packed_size != 0)) {
+ goto out_bad;
}
- launchd_assumes(launchd_mport_deallocate(mhp) == KERN_SUCCESS);
-}
+ launch_data_free(outdata_obj_array);
-void
-machservice_setup(launch_data_t obj, const char *key, void *context)
-{
- struct jobcb *j = context;
- struct machservice *ms;
- mach_port_t p = MACH_PORT_NULL;
+ *portsp = ports;
+ *ports_cnt = cnt;
- if ((ms = job_lookup_service(j->parent, key, false))) {
- job_log(j, LOG_WARNING, "Conflict with job: %s over Mach service: %s", ms->job->label, key);
- return;
- }
+ *reqport = jm->req_port;
+ *rcvright = jm->jm_port;
- if ((ms = machservice_new(j, key, &p)) == NULL) {
- job_log_error(j, LOG_WARNING, "Cannot add service: %s", key);
- return;
- }
+ jm->req_port = 0;
+ jm->jm_port = 0;
- ms->isActive = false;
+ workaround_5477111 = j;
- if (launch_data_get_type(obj) == LAUNCH_DATA_DICTIONARY) {
- launch_data_dict_iterate(obj, machservice_setup_options, ms);
+ jobmgr_shutdown(jm);
+
+ return BOOTSTRAP_SUCCESS;
+
+out_bad:
+ if (outdata_obj_array) {
+ launch_data_free(outdata_obj_array);
+ }
+ if (*outdata) {
+ mig_deallocate(*outdata, *outdataCnt);
+ }
+ if (ports) {
+ mig_deallocate((vm_address_t)ports, cnt * sizeof(ports[0]));
}
-}
-struct jobcb *
-job_parent(struct jobcb *j)
-{
- return j->parent;
+ return BOOTSTRAP_NO_MEMORY;
}
-void
-job_uncork_fork(struct jobcb *j)
+kern_return_t
+job_mig_subset(job_t j, mach_port_t requestorport, mach_port_t *subsetportp)
{
- pid_t c = j->p;
+ int bsdepth = 0;
+ jobmgr_t jmr;
- if (j->stall_before_exec) {
- job_log(j, LOG_DEBUG, "Uncorking the fork().");
- /* this unblocks the child and avoids a race
- * between the above fork() and the kevent_mod() */
- write(j->execfd, &c, sizeof(c));
- j->stall_before_exec = false;
- } else {
- job_log(j, LOG_WARNING, "Attempt to uncork a job that isn't in the middle of a fork().");
+ if (!launchd_assumes(j != NULL)) {
+ return BOOTSTRAP_NO_MEMORY;
}
-}
-void
-job_foreach_service(struct jobcb *j, void (*bs_iter)(struct machservice *, void *), void *context, bool include_subjobs)
-{
- struct machservice *ms;
- struct jobcb *ji;
+ jmr = j->mgr;
- j = job_get_bs(j);
+ while ((jmr = jobmgr_parent(jmr)) != NULL) {
+ bsdepth++;
+ }
- if (include_subjobs) {
- SLIST_FOREACH(ji, &j->jobs, sle) {
- if (ji->req_port)
- continue;
+ /* Since we use recursion, we need an artificial depth for subsets */
+ if (unlikely(bsdepth > 100)) {
+ job_log(j, LOG_ERR, "Mach sub-bootstrap create request failed. Depth greater than: %d", bsdepth);
+ return BOOTSTRAP_NO_MEMORY;
+ }
+
+ char name[NAME_MAX];
+ snprintf(name, sizeof(name), "%s[%i].subset.%i", j->anonymous ? j->prog : j->label, j->p, MACH_PORT_INDEX(requestorport));
- SLIST_FOREACH(ms, &ji->machservices, sle)
- bs_iter(ms, context);
+ if (!job_assumes(j, (jmr = jobmgr_new(j->mgr, requestorport, MACH_PORT_NULL, false, name, true, j->asport)) != NULL)) {
+ if (unlikely(requestorport == MACH_PORT_NULL)) {
+ return BOOTSTRAP_NOT_PRIVILEGED;
}
+ return BOOTSTRAP_NO_MEMORY;
}
- SLIST_FOREACH(ms, &j->machservices, sle)
- bs_iter(ms, context);
+ *subsetportp = jmr->jm_port;
+ jmr->properties |= BOOTSTRAP_PROPERTY_EXPLICITSUBSET;
+
+ /* A job could create multiple subsets, so only add a reference the first time
+ * it does so we don't have to keep a count.
+ */
+ if (j->anonymous && !j->holds_ref) {
+ j->holds_ref = true;
+ runtime_add_ref();
+ }
+
+ job_log(j, LOG_DEBUG, "Job created a subset named \"%s\"", jmr->name);
+ return BOOTSTRAP_SUCCESS;
}
-struct jobcb *
-job_new_bootstrap(struct jobcb *p, mach_port_t requestorport, mach_port_t checkin_port)
+#ifndef __LAUNCH_DISABLE_XPC_SUPPORT__
+job_t
+xpc_domain_import_service(jobmgr_t jm, launch_data_t pload)
{
- char bslabel[1024] = "100000";
- struct jobcb *j;
+ jobmgr_t where2put = NULL;
+
+ launch_data_t destname = launch_data_dict_lookup(pload, LAUNCH_JOBKEY_XPCDOMAIN);
+ if (destname) {
+ if (launch_data_get_type(destname) == LAUNCH_DATA_STRING) {
+ const char *str = launch_data_get_string(destname);
+ if (strcmp(str, XPC_DOMAIN_TYPE_SYSTEM) == 0) {
+ where2put = _s_xpc_system_domain;
+ } else if (strcmp(str, XPC_DOMAIN_TYPE_PERUSER) == 0) {
+ where2put = jobmgr_find_xpc_per_user_domain(jm, jm->req_euid);
+ } else if (strcmp(str, XPC_DOMAIN_TYPE_PERSESSION) == 0) {
+ where2put = jobmgr_find_xpc_per_session_domain(jm, jm->req_asid);
+ } else {
+ jobmgr_log(jm, LOG_ERR, "Invalid XPC domain type: %s", str);
+ errno = EINVAL;
+ }
+ } else {
+ jobmgr_log(jm, LOG_ERR, "XPC domain type is not a string.");
+ errno = EINVAL;
+ }
- if (requestorport == MACH_PORT_NULL) {
- if (p) {
- job_log(p, LOG_ERR, "Mach sub-bootstrap create request requires a requester port");
+ if (where2put) {
+ launch_data_t mi = NULL;
+ if ((mi = launch_data_dict_lookup(pload, LAUNCH_JOBKEY_MULTIPLEINSTANCES))) {
+ if (launch_data_get_type(mi) == LAUNCH_DATA_BOOL && launch_data_get_bool(mi)) {
+ jobmgr_log(where2put, LOG_ERR, "Multiple-instance services are not supported in this domain.");
+ where2put = NULL;
+ errno = EINVAL;
+ }
+ }
}
- return NULL;
+ } else {
+ where2put = jm;
}
- j = job_new(p, bslabel, NULL, NULL, NULL, requestorport);
-
- if (j == NULL)
- return NULL;
-
- if (checkin_port != MACH_PORT_NULL) {
- j->bs_port = checkin_port;
- } else if (!launchd_assumes(launchd_mport_create_recv(&j->bs_port) == KERN_SUCCESS)) {
- goto out_bad;
+ job_t j = NULL;
+ if (where2put) {
+ jobmgr_log(where2put, LOG_DEBUG, "Importing service...");
+ j = jobmgr_import2(where2put, pload);
+ if (j) {
+ j->xpc_service = true;
+ if (where2put->xpc_singleton) {
+ /* If the service was destined for one of the global domains,
+ * then we have to alias it into our local domain to reserve the
+ * name.
+ */
+ job_t ja = job_new_alias(jm, j);
+ if (!ja) {
+ /* If we failed to alias the job because of a conflict over
+ * the label, then we remove it from the global domain. We
+ * don't want to risk having imported a malicious job into
+ * one of the global domains.
+ */
+ if (errno != EEXIST) {
+ job_assumes(j, errno == 0);
+ } else {
+ job_log(j, LOG_ERR, "Failed to alias job into: %s", where2put->name);
+ }
+
+ job_remove(j);
+ } else {
+ ja->xpc_service = true;
+ j = ja;
+ }
+ }
+ }
}
- sprintf(j->label, "%d", MACH_PORT_INDEX(j->bs_port));
-
- if (!launchd_assumes(launchd_mport_request_callback(j->bs_port, j, true) == KERN_SUCCESS))
- goto out_bad;
+ return j;
+}
- if (p) {
- job_log(p, LOG_DEBUG, "Mach sub-bootstrap created: %s", j->label);
+kern_return_t
+xpc_domain_import2(job_t j, mach_port_t reqport, mach_port_t dport)
+{
+ if (unlikely(!pid1_magic)) {
+ job_log(j, LOG_ERR, "XPC domains may only reside in PID 1.");
+ return BOOTSTRAP_NOT_PRIVILEGED;
+ }
+ if (!MACH_PORT_VALID(reqport)) {
+ return BOOTSTRAP_UNKNOWN_SERVICE;
}
- return j;
+ kern_return_t kr = BOOTSTRAP_NO_MEMORY;
+ /* All XPC domains are children of the root job manager. What we're creating
+ * here is really just a skeleton. By creating it, we're adding reqp to our
+ * port set. It will have two messages on it. The first specifies the
+ * environment of the originator. This is so we can cache it and hand it to
+ * xpcproxy to bootstrap our services. The second is the set of jobs that is
+ * to be bootstrapped in.
+ */
+ jobmgr_t jm = jobmgr_new(root_jobmgr, reqport, dport, false, NULL, true, MACH_PORT_NULL);
+ if (job_assumes(j, jm != NULL)) {
+ jm->properties |= BOOTSTRAP_PROPERTY_XPC_DOMAIN;
+ jm->shortdesc = "private";
+ kr = BOOTSTRAP_SUCCESS;
+ }
-out_bad:
- if (j)
- job_remove(j);
- return NULL;
+ return kr;
}
-void
-job_delete_anything_with_port(struct jobcb *j, mach_port_t port)
+kern_return_t
+xpc_domain_set_environment(job_t j, mach_port_t rp, mach_port_t bsport, mach_port_t excport, vm_offset_t ctx, mach_msg_type_number_t ctx_sz)
{
- struct machservice *ms, *next_ms;
- struct jobcb *ji, *jn;
-
- /* Mach ports, unlike Unix descriptors, are reference counted. In other
- * words, when some program hands us a second or subsequent send right
- * to a port we already have open, the Mach kernel gives us the same
- * port number back and increments an reference count associated with
- * the port. This forces us, when discovering that a receive right at
- * the other end has been deleted, to wander all of our objects to see
- * what weird places clients might have handed us the same send right
- * to use.
- */
+ if (!j) {
+ /* Due to the whacky nature of XPC service bootstrapping, we can end up
+ * getting this message long after the requesting process has gone away.
+ * See <rdar://problem/8593143>.
+ */
+ return BOOTSTRAP_UNKNOWN_SERVICE;
+ }
- if (j->req_port == port)
- return job_remove(j);
+ jobmgr_t jm = j->mgr;
+ if (!(jm->properties & BOOTSTRAP_PROPERTY_XPC_DOMAIN)) {
+ return BOOTSTRAP_NOT_PRIVILEGED;
+ }
+
+ if (jm->req_asport != MACH_PORT_NULL) {
+ return BOOTSTRAP_NOT_PRIVILEGED;
+ }
+
+ struct ldcred *ldc = runtime_get_caller_creds();
+ struct proc_bsdshortinfo proc;
+ if (proc_pidinfo(ldc->pid, PROC_PIDT_SHORTBSDINFO, 1, &proc, PROC_PIDT_SHORTBSDINFO_SIZE) == 0) {
+ if (errno != ESRCH) {
+ jobmgr_assumes(jm, errno == 0);
+ }
- SLIST_FOREACH_SAFE(ji, &j->jobs, sle, jn)
- job_delete_anything_with_port(ji, port);
+ jm->error = errno;
+ jobmgr_remove(jm);
+ return BOOTSTRAP_NO_MEMORY;
+ }
- SLIST_FOREACH_SAFE(ms, &j->machservices, sle, next_ms) {
- if (ms->port == port)
- machservice_delete(ms);
+ if (!jobmgr_assumes(jm, audit_session_port(ldc->asid, &jm->req_asport) == 0)) {
+ jm->error = EPERM;
+ jobmgr_remove(jm);
+ job_log(j, LOG_ERR, "Failed to get port for ASID: %u", ldc->asid);
+ return BOOTSTRAP_NOT_PRIVILEGED;
}
+
+ (void)snprintf(jm->name_init, NAME_MAX, "com.apple.xpc.domain.%s[%i]", proc.pbsi_comm, ldc->pid);
+ strlcpy(jm->owner, proc.pbsi_comm, sizeof(jm->owner));
+ jm->req_bsport = bsport;
+ jm->req_excport = excport;
+ jm->req_rport = rp;
+ jm->req_ctx = ctx;
+ jm->req_ctx_sz = ctx_sz;
+ jm->req_pid = ldc->pid;
+ jm->req_euid = ldc->euid;
+ jm->req_egid = ldc->egid;
+ jm->req_asid = ldc->asid;
+
+ return KERN_SUCCESS;
}
-struct machservice *
-job_lookup_service(struct jobcb *j, const char *name, bool check_parent)
+kern_return_t
+xpc_domain_load_services(job_t j, vm_offset_t services_buff, mach_msg_type_number_t services_sz)
{
- struct machservice *ms;
- struct jobcb *ji;
+ if (!j) {
+ return BOOTSTRAP_UNKNOWN_SERVICE;
+ }
+
+ /* This is just for XPC domains (for now). */
+ if (!(j->mgr->properties & BOOTSTRAP_PROPERTY_XPC_DOMAIN)) {
+ return BOOTSTRAP_NOT_PRIVILEGED;
+ }
+ if (j->mgr->session_initialized) {
+ jobmgr_log(j->mgr, LOG_ERR, "Attempt to initialize an already-initialized XPC domain.");
+ return BOOTSTRAP_NOT_PRIVILEGED;
+ }
- j = job_get_bs(j);
+ size_t offset = 0;
+ launch_data_t services = launch_data_unpack((void *)services_buff, services_sz, NULL, 0, &offset, NULL);
+ if (!jobmgr_assumes(j->mgr, services != NULL)) {
+ return BOOTSTRAP_NO_MEMORY;
+ }
- SLIST_FOREACH(ji, &j->jobs, sle) {
- if (ji->req_port)
- continue;
+ size_t i = 0;
+ size_t c = launch_data_array_get_count(services);
+ for (i = 0; i < c; i++) {
+ job_t nj = NULL;
+ launch_data_t ploadi = launch_data_array_get_index(services, i);
+ if (!(nj = xpc_domain_import_service(j->mgr, ploadi))) {
+ /* If loading one job fails, just fail the whole thing. At this
+ * point, xpchelper should receive the failure and then just refuse
+ * to launch the application, since its XPC services could not be
+ * fully bootstrapped.
+ *
+ * Take care to not reference the job or its manager after this
+ * point.
+ */
+ if (errno == EINVAL) {
+ jobmgr_log(j->mgr, LOG_ERR, "Service at index is not valid: %lu", i);
+ } else if (errno == EEXIST) {
+ /* If we get back EEXIST, we know that the payload was a
+ * dictionary with a label. But, well, I guess it never hurts to
+ * check.
+ */
+ char *label = "(bogus)";
+ if (launch_data_get_type(ploadi) == LAUNCH_DATA_DICTIONARY) {
+ launch_data_t llabel = launch_data_dict_lookup(ploadi, LAUNCH_JOBKEY_LABEL);
+ if (launch_data_get_type(llabel) == LAUNCH_DATA_STRING) {
+ label = (char *)launch_data_get_string(llabel);
+ }
+ }
+ jobmgr_log(j->mgr, LOG_ERR, "Service name conflict: %s", label);
+ }
- SLIST_FOREACH(ms, &ji->machservices, sle) {
- if (strcmp(name, ms->name) == 0)
- return ms;
+ j->mgr->error = errno;
+ jobmgr_log(j->mgr, LOG_ERR, "Obliterating domain.");
+ jobmgr_remove(j->mgr);
+ break;
+ } else {
+ jobmgr_log(j->mgr, LOG_DEBUG, "Imported service %s", nj->label);
+ job_dispatch(nj, false);
}
}
- SLIST_FOREACH(ms, &j->machservices, sle) {
- if (strcmp(name, ms->name) == 0)
- return ms;
+ kern_return_t result = BOOTSTRAP_NO_MEMORY;
+ if (i == c) {
+ j->mgr->session_initialized = true;
+ (void)jobmgr_assumes(j->mgr, xpc_call_wakeup(j->mgr->req_rport, BOOTSTRAP_SUCCESS) == KERN_SUCCESS);
+ j->mgr->req_rport = MACH_PORT_NULL;
+
+ /* Returning a failure code will destroy the message, whereas returning
+ * success will not, so we need to clean up here.
+ */
+ mig_deallocate(services_buff, services_sz);
+ result = BOOTSTRAP_SUCCESS;
}
- if (j->parent == NULL)
- return NULL;
-
- if (!check_parent)
- return NULL;
-
- return job_lookup_service(j->parent, name, true);
+ return result;
}
-mach_port_t
-machservice_port(struct machservice *ms)
+kern_return_t
+xpc_domain_check_in(job_t j, mach_port_t *bsport, mach_port_t *sbsport, mach_port_t *excport, mach_port_t *asport, uint32_t *uid, uint32_t *gid, int32_t *asid, vm_offset_t *ctx, mach_msg_type_number_t *ctx_sz)
{
- return ms->port;
+ if (!jobmgr_assumes(root_jobmgr, j != NULL)) {
+ return BOOTSTRAP_UNKNOWN_SERVICE;
+ }
+ jobmgr_t jm = j->mgr;
+ if (!(jm->properties & BOOTSTRAP_PROPERTY_XPC_DOMAIN)) {
+ return BOOTSTRAP_NOT_PRIVILEGED;
+ }
+
+ if (jm->req_asport == MACH_PORT_NULL) {
+ return BOOTSTRAP_NOT_PRIVILEGED;
+ }
+
+ *bsport = jm->req_bsport;
+ *sbsport = root_jobmgr->jm_port;
+ *excport = jm->req_excport;
+ *asport = jm->req_asport;
+ *uid = jm->req_euid;
+ *gid = jm->req_egid;
+ *asid = jm->req_asid;
+
+ *ctx = jm->req_ctx;
+ *ctx_sz = jm->req_ctx_sz;
+
+ return KERN_SUCCESS;
}
-struct jobcb *
-machservice_job(struct machservice *ms)
+kern_return_t
+xpc_domain_get_service_name(job_t j, event_name_t name)
{
- return ms->job;
-}
+ if (!j) {
+ return BOOTSTRAP_NO_MEMORY;
+ }
+ if (!j->xpc_service) {
+ jobmgr_log(j->mgr, LOG_ERR, "Attempt to get service name by non-XPC service: %s", j->label);
+ return BOOTSTRAP_NOT_PRIVILEGED;
+ }
-bool
-machservice_hidden(struct machservice *ms)
-{
- return ms->hide;
+ struct machservice * ms = SLIST_FIRST(&j->machservices);
+ if (!ms) {
+ jobmgr_log(j->mgr, LOG_ERR, "Attempt to get service name of job with no machservices: %s", j->label);
+ return BOOTSTRAP_UNKNOWN_SERVICE;
+ }
+
+ (void)strlcpy(name, ms->name, sizeof(event_name_t));
+ return BOOTSTRAP_SUCCESS;
}
+#endif
-bool
-machservice_active(struct machservice *ms)
+kern_return_t
+xpc_events_get_channel_name(job_t j __attribute__((unused)), event_name_t stream __attribute__((unused)), uint64_t token __attribute__((unused)), event_name_t name __attribute__((unused)))
{
- return ms->isActive;
+ return KERN_FAILURE;
}
-const char *
-machservice_name(struct machservice *ms)
+kern_return_t
+xpc_events_get_event_name(job_t j, event_name_t stream, uint64_t token, event_name_t name)
{
- return ms->name;
-}
+ struct externalevent *event = externalevent_find(stream, token);
+ if (event && j->event_monitor) {
+ (void)strcpy(name, event->name);
+ } else {
+ event = NULL;
+ }
-void
-machservice_delete(struct machservice *ms)
+ return event ? BOOTSTRAP_SUCCESS : BOOTSTRAP_UNKNOWN_SERVICE;
+}
+
+kern_return_t
+xpc_events_set_event(job_t j, event_name_t stream, event_name_t key, vm_offset_t event, mach_msg_type_number_t eventCnt)
{
- if (ms->recv) {
- if (ms->isActive) {
- /* FIXME we should cancel the notification */
- } else {
- launchd_assumes(launchd_mport_close_recv(ms->port) == KERN_SUCCESS);
+ if (j->anonymous) {
+ return BOOTSTRAP_NOT_PRIVILEGED;
+ }
+
+ struct externalevent *eei = NULL;
+ LIST_FOREACH(eei, &j->events, job_le) {
+ if (strcmp(eei->name, key) == 0 && strcmp(eei->sys->name, stream) == 0) {
+ externalevent_delete(eei);
+ eventsystem_ping();
+ break;
}
}
- launchd_assumes(launchd_mport_deallocate(ms->port) == KERN_SUCCESS);
+ bool success = false;
+ struct eventsystem *es = eventsystem_find(stream);
+ if (!es) {
+ es = eventsystem_new(stream);
+ (void)job_assumes(j, es != NULL);
+ }
- job_log(ms->job, LOG_INFO, "Mach service deleted: %s", ms->name);
+ if (es) {
+ size_t offset = 0;
+ launch_data_t unpacked = launch_data_unpack((void *)event, eventCnt, NULL, 0, &offset, 0);
+ if (unpacked && launch_data_get_type(unpacked) == LAUNCH_DATA_DICTIONARY) {
+ success = externalevent_new(j, es, key, unpacked);
+ }
+ }
- SLIST_REMOVE(&ms->job->machservices, ms, machservice, sle);
+ if (!success) {
+ mig_deallocate(event, eventCnt);
+ }
- free(ms);
+ return KERN_SUCCESS;
}
-void
-machservice_watch(struct machservice *ms)
+kern_return_t
+xpc_events_get_event(job_t j, event_name_t stream, event_name_t key, vm_offset_t *event, mach_msg_type_number_t *eventCnt)
{
- mach_msg_id_t which = MACH_NOTIFY_DEAD_NAME;
-
- ms->isActive = true;
+ struct externalevent *eei = NULL;
+ LIST_FOREACH(eei, &j->events, job_le) {
+ if (strcmp(eei->name, key) == 0 && strcmp(eei->sys->name, stream) == 0) {
+ /* Big enough. */
+ *eventCnt = 10 * 1024;
+ mig_allocate(event, *eventCnt);
+
+ size_t sz = launch_data_pack(eei->event, (void *)*event, *eventCnt, NULL, NULL);
+ if (!job_assumes(j, sz != 0)) {
+ mig_deallocate(*event, *eventCnt);
+ return BOOTSTRAP_NO_MEMORY;
+ }
- if (ms->job->req_port == MACH_PORT_NULL) {
- which = MACH_NOTIFY_PORT_DESTROYED;
- job_checkin(ms->job);
+ return BOOTSTRAP_SUCCESS;
+ }
}
- launchd_assumes(launchd_mport_notify_req(ms->port, which) == KERN_SUCCESS);
+ return BOOTSTRAP_UNKNOWN_SERVICE;
}
-#define NELEM(x) (sizeof(x)/sizeof(x[0]))
-#define END_OF(x) (&(x)[NELEM(x)])
-
-char **
-mach_cmd2argv(const char *string)
+struct machservice *
+xpc_events_find_channel(job_t j, event_name_t stream, mach_port_t *p)
{
- char *argv[100], args[1000];
- const char *cp;
- char *argp = args, term, **argv_ret, *co;
- unsigned int nargs = 0, i;
-
- for (cp = string; *cp;) {
- while (isspace(*cp))
- cp++;
- term = (*cp == '"') ? *cp++ : '\0';
- if (nargs < NELEM(argv))
- argv[nargs++] = argp;
- while (*cp && (term ? *cp != term : !isspace(*cp)) && argp < END_OF(args)) {
- if (*cp == '\\')
- cp++;
- *argp++ = *cp;
- if (*cp)
- cp++;
+ struct machservice *msi = NULL;
+ SLIST_FOREACH(msi, &j->machservices, sle) {
+ if (strcmp(stream, msi->name) == 0) {
+ break;
}
- *argp++ = '\0';
}
- argv[nargs] = NULL;
-
- if (nargs == 0)
- return NULL;
-
- argv_ret = malloc((nargs + 1) * sizeof(char *) + strlen(string) + 1);
-
- if (!launchd_assumes(argv_ret != NULL))
- return NULL;
-
- co = (char *)argv_ret + (nargs + 1) * sizeof(char *);
- for (i = 0; i < nargs; i++) {
- strcpy(co, argv[i]);
- argv_ret[i] = co;
- co += strlen(argv[i]) + 1;
+ if (!msi) {
+ mach_port_t sp = MACH_PORT_NULL;
+ msi = machservice_new(j, stream, &sp, false);
+ if (job_assumes(j, msi)) {
+ /* Hack to keep this from being publicly accessible through
+ * bootstrap_look_up().
+ */
+ LIST_REMOVE(msi, name_hash_sle);
+ msi->event_channel = true;
+ *p = sp;
+
+ machservice_watch(j, msi);
+ } else {
+ errno = BOOTSTRAP_NO_MEMORY;
+ }
+ } else {
+ if (!msi->event_channel) {
+ job_log(j, LOG_ERR, "This job registered a MachService name identical to the requested event channel name: %s", stream);
+ msi = NULL;
+ errno = BOOTSTRAP_NAME_IN_USE;
+ } else {
+ *p = msi->port;
+ }
}
- argv_ret[i] = NULL;
-
- return argv_ret;
+
+ return msi;
}
-void
-job_checkin(struct jobcb *j)
+kern_return_t
+xpc_events_channel_check_in(job_t j, event_name_t stream, uint64_t flags __attribute__((unused)), mach_port_t *p)
{
- j->checkedin = true;
+ struct machservice *ms = xpc_events_find_channel(j, stream, p);
+ if (ms) {
+ if (ms->isActive) {
+ job_log(j, LOG_ERR, "Attempt to check in on event channel multiple times: %s", stream);
+ *p = MACH_PORT_NULL;
+ errno = BOOTSTRAP_SERVICE_ACTIVE;
+ } else {
+ job_checkin(j);
+ machservice_request_notifications(ms);
+ errno = BOOTSTRAP_SUCCESS;
+ }
+ }
+
+ return errno;
}
-bool
-job_ack_port_destruction(struct jobcb *j, mach_port_t p)
+kern_return_t
+xpc_events_channel_look_up(job_t j, event_name_t stream, event_token_t token, uint64_t flags __attribute__((unused)), mach_port_t *p)
{
- struct jobcb *ji;
- struct machservice *ms;
+ if (!j->event_monitor) {
+ return BOOTSTRAP_NOT_PRIVILEGED;
+ }
- SLIST_FOREACH(ji, &j->jobs, sle) {
- if (job_ack_port_destruction(ji, p))
- return true;
+ struct externalevent *ee = externalevent_find(stream, token);
+ if (!ee) {
+ return BOOTSTRAP_UNKNOWN_SERVICE;
}
- SLIST_FOREACH(ms, &j->machservices, sle) {
- if (ms->port == p)
- break;
+ struct machservice *ms = xpc_events_find_channel(ee->job, stream, p);
+ if (ms) {
+ errno = BOOTSTRAP_SUCCESS;
}
- if (ms == NULL)
- return false;
+ return errno;
+}
- ms->isActive = false;
+kern_return_t
+job_mig_kickstart(job_t j, name_t targetlabel, pid_t *out_pid, unsigned int flags)
+{
+ struct ldcred *ldc = runtime_get_caller_creds();
+ job_t otherj;
- if (ms->reset)
- machservice_resetport(j, ms);
+ if (!launchd_assumes(j != NULL)) {
+ return BOOTSTRAP_NO_MEMORY;
+ }
- job_log(j, LOG_DEBUG, "Receive right returned to us: %s", ms->name);
+ if (unlikely(!(otherj = job_find(NULL, targetlabel)))) {
+ return BOOTSTRAP_UNKNOWN_SERVICE;
+ }
- job_dispatch(j);
+#if TARGET_OS_EMBEDDED
+ bool allow_non_root_kickstart = j->username && otherj->username && (strcmp(j->username, otherj->username) == 0);
+#else
+ bool allow_non_root_kickstart = false;
+#endif
- return true;
-}
+ if (ldc->euid != 0 && ldc->euid != geteuid() && !allow_non_root_kickstart) {
+ return BOOTSTRAP_NOT_PRIVILEGED;
+ }
-void
-job_ack_no_senders(struct jobcb *j)
-{
- j->priv_port_has_senders = false;
+#if HAVE_SANDBOX
+ if (unlikely(sandbox_check(ldc->pid, "job-creation", SANDBOX_FILTER_NONE) > 0)) {
+ return BOOTSTRAP_NOT_PRIVILEGED;
+ }
+#endif
- job_log(j, LOG_DEBUG, "No more senders on privileged Mach bootstrap port");
+ if (otherj->p && (flags & VPROCFLAG_STALL_JOB_EXEC)) {
+ return BOOTSTRAP_SERVICE_ACTIVE;
+ }
- job_dispatch(j);
-}
+ otherj->stall_before_exec = (flags & VPROCFLAG_STALL_JOB_EXEC);
+ otherj = job_dispatch(otherj, true);
-mach_port_t
-job_get_reqport(struct jobcb *j)
-{
- j->transfer_bstrap = true;
- gc_this_job = j;
+ if (!job_assumes(j, otherj && otherj->p)) {
+ /* <rdar://problem/6787083> Clear this flag if we failed to start the job. */
+ otherj->stall_before_exec = false;
+ return BOOTSTRAP_NO_MEMORY;
+ }
- return j->req_port;
-}
+ *out_pid = otherj->p;
-mach_port_t
-job_get_bsport(struct jobcb *j)
-{
- return j->bs_port;
+ return 0;
}
-struct jobcb *
-job_get_bs(struct jobcb *j)
+kern_return_t
+job_mig_spawn_internal(job_t j, vm_offset_t indata, mach_msg_type_number_t indataCnt, mach_port_t asport, job_t *outj)
{
- if (j->req_port)
- return j;
+ launch_data_t jobdata = NULL;
+ size_t data_offset = 0;
+ struct ldcred *ldc = runtime_get_caller_creds();
+ job_t jr;
+
+ if (!launchd_assumes(j != NULL)) {
+ return BOOTSTRAP_NO_MEMORY;
+ }
- if (launchd_assumes(j->parent != NULL))
- return j->parent;
+ if (unlikely(j->deny_job_creation)) {
+ return BOOTSTRAP_NOT_PRIVILEGED;
+ }
- return NULL;
-}
+#if HAVE_SANDBOX
+ if (unlikely(sandbox_check(ldc->pid, "job-creation", SANDBOX_FILTER_NONE) > 0)) {
+ return BOOTSTRAP_NOT_PRIVILEGED;
+ }
+#endif
+
+ if (unlikely(pid1_magic && ldc->euid && ldc->uid)) {
+ job_log(j, LOG_DEBUG, "Punting spawn to per-user-context");
+ return VPROC_ERR_TRY_PER_USER;
+ }
-pid_t
-job_get_pid(struct jobcb *j)
-{
- return j->p;
-}
+ if (!job_assumes(j, indataCnt != 0)) {
+ return 1;
+ }
-bool
-semaphoreitem_new(struct jobcb *j, semaphore_reason_t why, const char *what)
-{
- struct semaphoreitem *si;
- size_t alloc_sz = sizeof(struct semaphoreitem);
+ runtime_ktrace0(RTKT_LAUNCHD_DATA_UNPACK);
+ if (!job_assumes(j, (jobdata = launch_data_unpack((void *)indata, indataCnt, NULL, 0, &data_offset, NULL)) != NULL)) {
+ return 1;
+ }
- if (what)
- alloc_sz += strlen(what) + 1;
+ jobmgr_t target_jm = jobmgr_find_by_name(j->mgr, NULL);
+ if (!jobmgr_assumes(j->mgr, target_jm != NULL)) {
+ jobmgr_log(j->mgr, LOG_ERR, "This API can only be used by a process running within an Aqua session.");
+ return 1;
+ }
- if (!launchd_assumes(si = calloc(1, alloc_sz)))
- return false;
+ jr = jobmgr_import2(target_jm ?: j->mgr, jobdata);
+
+ launch_data_t label = NULL;
+ launch_data_t wait4debugger = NULL;
+ if (!jr) {
+ switch (errno) {
+ case EEXIST:
+ /* If EEXIST was returned, we know that there is a label string in
+ * the dictionary. So we don't need to check the types here; that
+ * has already been done.
+ */
+ label = launch_data_dict_lookup(jobdata, LAUNCH_JOBKEY_LABEL);
+ jr = job_find(NULL, launch_data_get_string(label));
+ if (job_assumes(j, jr != NULL) && !jr->p) {
+ wait4debugger = launch_data_dict_lookup(jobdata, LAUNCH_JOBKEY_WAITFORDEBUGGER);
+ if (wait4debugger && launch_data_get_type(wait4debugger) == LAUNCH_DATA_BOOL) {
+ if (launch_data_get_bool(wait4debugger)) {
+ /* If the job exists, we're going to kick-start it, but
+ * we need to give the caller the opportunity to start
+ * it suspended if it so desires. But this will only
+ * take effect if the job isn't running.
+ */
+ jr->wait4debugger_oneshot = true;
+ }
+ }
+ }
- si->why = why;
+ *outj = jr;
+ return BOOTSTRAP_NAME_IN_USE;
+ default:
+ return BOOTSTRAP_NO_MEMORY;
+ }
+ }
- if (what)
- strcpy(si->what, what);
+ if (pid1_magic) {
+ jr->mach_uid = ldc->uid;
+ }
- SLIST_INSERT_HEAD(&j->semaphores, si, sle);
+ jr->legacy_LS_job = true;
+ jr->abandon_pg = true;
+ jr->asport = asport;
+ uuid_clear(jr->expected_audit_uuid);
+ jr = job_dispatch(jr, true);
- return true;
+ if (!job_assumes(j, jr != NULL)) {
+ job_remove(jr);
+ return BOOTSTRAP_NO_MEMORY;
+ }
+
+ if (!job_assumes(jr, jr->p)) {
+ job_remove(jr);
+ return BOOTSTRAP_NO_MEMORY;
+ }
+
+ job_log(jr, LOG_DEBUG, "Spawned by PID %u: %s", j->p, j->label);
+ *outj = jr;
+
+ return BOOTSTRAP_SUCCESS;
}
-void
-semaphoreitem_delete(struct jobcb *j, struct semaphoreitem *ri)
+kern_return_t
+job_mig_spawn2(job_t j, mach_port_t rp, vm_offset_t indata, mach_msg_type_number_t indataCnt, mach_port_t asport, pid_t *child_pid, mach_port_t *obsvr_port)
{
- SLIST_REMOVE(&j->semaphores, ri, semaphoreitem, sle);
+ job_t nj = NULL;
+ kern_return_t kr = job_mig_spawn_internal(j, indata, indataCnt, asport, &nj);
+ if (likely(kr == KERN_SUCCESS)) {
+ if (job_setup_exit_port(nj) != KERN_SUCCESS) {
+ job_remove(nj);
+ kr = BOOTSTRAP_NO_MEMORY;
+ } else {
+ /* Do not return until the job has called exec(3), thereby making it
+ * safe for the caller to send it SIGCONT.
+ *
+ * <rdar://problem/9042798>
+ */
+ nj->spawn_reply_port = rp;
+ kr = MIG_NO_REPLY;
+ }
+ } else if (kr == BOOTSTRAP_NAME_IN_USE) {
+ bool was_running = nj->p;
+ if (job_dispatch(nj, true)) {
+ if (!was_running) {
+ job_log(nj, LOG_DEBUG, "Job exists but is not running. Kick-starting.");
+
+ if (job_setup_exit_port(nj) == KERN_SUCCESS) {
+ nj->spawn_reply_port = rp;
+ kr = MIG_NO_REPLY;
+ } else {
+ kr = BOOTSTRAP_NO_MEMORY;
+ }
+ } else {
+ *obsvr_port = MACH_PORT_NULL;
+ *child_pid = nj->p;
+ kr = KERN_SUCCESS;
+ }
+ } else {
+ job_log(nj, LOG_ERR, "Failed to dispatch job, requestor: %s", j->label);
+ kr = BOOTSTRAP_UNKNOWN_SERVICE;
+ }
+ }
- free(ri);
+ mig_deallocate(indata, indataCnt);
+ return kr;
}
-void
-semaphoreitem_setup_paths(launch_data_t obj, const char *key, void *context)
+kern_return_t
+job_mig_event_source_check_in(job_t j, name_t name, mach_port_t ping_port, vm_offset_t *outval, mach_msg_type_number_t *outvalCnt, uint64_t *tokens)
{
- struct jobcb *j = context;
- semaphore_reason_t why;
+ if (!j || !j->event_monitor) {
+ return BOOTSTRAP_NOT_PRIVILEGED;
+ }
- why = launch_data_get_bool(obj) ? PATH_EXISTS : PATH_MISSING;
+ /* Update our ping-port. One ping will force all the notification systems
+ * to check in, so they'll all give us send-once rights. It doesn't really
+ * matter which one we keep around. It's not the most efficient thing ever,
+ * but keep in mind that, by doing this over one channel, we can do it over
+ * the job's MachService. This means that we'll get it back when the job dies,
+ * and we can create ourselves a send-once right if we didn't have one already,
+ * and we can just keep the helper alive without it needing to bootstrap
+ * communication.
+ *
+ * So we're trading efficiency for robustness. In this case, the checkins
+ * should happen pretty infrequently, so it's pretty worth it.
+ */
+ if (_s_event_update_port != MACH_PORT_NULL) {
+ (void)job_assumes(j, launchd_mport_deallocate(_s_event_update_port) == KERN_SUCCESS);
+ }
+ _s_event_update_port = ping_port;
+
+ kern_return_t result = BOOTSTRAP_NO_MEMORY;
+ launch_data_t arr = launch_data_alloc(LAUNCH_DATA_ARRAY);
+ if (job_assumes(j, arr != NULL)) {
+ struct eventsystem *es = eventsystem_find(name);
+ if (unlikely(es == NULL)) {
+ es = eventsystem_new(name);
+ }
+
+ if (job_assumes(j, es != NULL)) {
+ struct externalevent *ei = NULL;
+ size_t i = 0;
+ LIST_FOREACH(ei, &es->events, sys_le) {
+ (void)job_assumes(j, launch_data_array_set_index(arr, ei->event, i));
+ if (job_assumes(j, i < 1024)) {
+ tokens[i] = ei->id;
+ } else {
+ break;
+ }
+ i++;
+ }
+
+ /* Big enough. */
+ *outvalCnt = 10 * 1024;
+ mig_allocate(outval, *outvalCnt);
+
+ size_t sz = launch_data_pack(arr, (void *)*outval, *outvalCnt, NULL, NULL);
+ if (job_assumes(j, sz != 0)) {
+ result = BOOTSTRAP_SUCCESS;
+ } else {
+ mig_deallocate(*outval, *outvalCnt);
+ }
+ }
- semaphoreitem_new(j, why, key);
+ /* Total hack, but launch_data doesn't do ref-counting. */
+ struct _launch_data *hack = (struct _launch_data *)arr;
+ free(hack->_array);
+ free(arr);
+ }
+
+ return result;
}
-void
-semaphoreitem_setup(launch_data_t obj, const char *key, void *context)
+kern_return_t
+job_mig_event_set_state(job_t j, name_t name, uint64_t token, boolean_t state)
{
- struct jobcb *j = context;
- semaphore_reason_t why;
-
- if (strcasecmp(key, LAUNCH_JOBKEY_KEEPALIVE_NETWORKSTATE) == 0) {
- why = launch_data_get_bool(obj) ? NETWORK_UP : NETWORK_DOWN;
- semaphoreitem_new(j, why, NULL);
- } else if (strcasecmp(key, LAUNCH_JOBKEY_KEEPALIVE_SUCCESSFULEXIT) == 0) {
- why = launch_data_get_bool(obj) ? SUCCESSFUL_EXIT : FAILED_EXIT;
- semaphoreitem_new(j, why, NULL);
- j->runatload = true;
- } else if (strcasecmp(key, LAUNCH_JOBKEY_KEEPALIVE_PATHSTATE) == 0 &&
- launch_data_get_type(obj) == LAUNCH_DATA_DICTIONARY) {
- launch_data_dict_iterate(obj, semaphoreitem_setup_paths, j);
+ if (!j->event_monitor) {
+ return BOOTSTRAP_NOT_PRIVILEGED;
+ }
+
+ struct externalevent *ei = externalevent_find(name, token);
+ if (job_assumes(j, ei != NULL)) {
+ ei->state = state;
+ if(job_dispatch(ei->job, false) == NULL) {
+ if (errno == EPERM) {
+ return BOOTSTRAP_NOT_PRIVILEGED;
+ }
+ return BOOTSTRAP_NO_MEMORY;
+ }
+ } else {
+ return BOOTSTRAP_NO_MEMORY;
}
+
+ return BOOTSTRAP_SUCCESS;
}
void
-job_dispatch_all_other_semaphores(struct jobcb *j, struct jobcb *nj)
+jobmgr_init(bool sflag)
{
- struct jobcb *ji;
-
- if (j == nj)
- return;
-
- if (!SLIST_EMPTY(&j->semaphores))
- job_dispatch(j);
+ const char *root_session_type = pid1_magic ? VPROCMGR_SESSION_SYSTEM : VPROCMGR_SESSION_BACKGROUND;
+ SLIST_INIT(&s_curious_jobs);
+ LIST_INIT(&s_needing_sessions);
+
+ launchd_assert((root_jobmgr = jobmgr_new(NULL, MACH_PORT_NULL, MACH_PORT_NULL, sflag, root_session_type, false, MACH_PORT_NULL)) != NULL);
+#ifndef __LAUNCH_DISABLE_XPC_SUPPORT__
+ launchd_assert((_s_xpc_system_domain = jobmgr_new_xpc_singleton_domain(root_jobmgr, "com.apple.xpc.system")) != NULL);
+ _s_xpc_system_domain->req_asid = g_audit_session;
+ _s_xpc_system_domain->req_asport = g_audit_session_port;
+ _s_xpc_system_domain->shortdesc = "system";
+#endif
+ if (pid1_magic) {
+ root_jobmgr->monitor_shutdown = true;
+ }
- SLIST_FOREACH(ji, &j->jobs, sle)
- job_dispatch_all_other_semaphores(ji, nj);
+ uint32_t fflags = NOTE_ATTRIB | NOTE_LINK | NOTE_REVOKE | NOTE_EXTEND | NOTE_WRITE;
+ s_no_hang_fd = open("/dev/autofs_nowait", O_EVTONLY | O_NONBLOCK);
+ if (likely(s_no_hang_fd == -1)) {
+ if (jobmgr_assumes(root_jobmgr, (s_no_hang_fd = open("/dev", O_EVTONLY | O_NONBLOCK)) != -1)) {
+ (void)jobmgr_assumes(root_jobmgr, kevent_mod((uintptr_t)s_no_hang_fd, EVFILT_VNODE, EV_ADD, fflags, 0, root_jobmgr) != -1);
+ }
+ }
+ s_no_hang_fd = _fd(s_no_hang_fd);
}
-time_t
-cronemu(int mon, int mday, int hour, int min)
+size_t
+our_strhash(const char *s)
{
- struct tm workingtm;
- time_t now;
-
- now = time(NULL);
- workingtm = *localtime(&now);
+ size_t c, r = 5381;
- workingtm.tm_isdst = -1;
- workingtm.tm_sec = 0;
- workingtm.tm_min++;
+ /* djb2
+ * This algorithm was first reported by Dan Bernstein many years ago in comp.lang.c
+ */
- while (!cronemu_mon(&workingtm, mon, mday, hour, min)) {
- workingtm.tm_year++;
- workingtm.tm_mon = 0;
- workingtm.tm_mday = 1;
- workingtm.tm_hour = 0;
- workingtm.tm_min = 0;
- mktime(&workingtm);
+ while ((c = *s++)) {
+ r = ((r << 5) + r) + c; /* hash*33 + c */
}
- return mktime(&workingtm);
+ return r;
}
-time_t
-cronemu_wday(int wday, int hour, int min)
+size_t
+hash_label(const char *label)
{
- struct tm workingtm;
- time_t now;
-
- now = time(NULL);
- workingtm = *localtime(&now);
-
- workingtm.tm_isdst = -1;
- workingtm.tm_sec = 0;
- workingtm.tm_min++;
-
- if (wday == 7)
- wday = 0;
-
- while (!(workingtm.tm_wday == wday && cronemu_hour(&workingtm, hour, min))) {
- workingtm.tm_mday++;
- workingtm.tm_hour = 0;
- workingtm.tm_min = 0;
- mktime(&workingtm);
- }
+ return our_strhash(label) % LABEL_HASH_SIZE;
+}
- return mktime(&workingtm);
+size_t
+hash_ms(const char *msstr)
+{
+ return our_strhash(msstr) % MACHSERVICE_HASH_SIZE;
}
bool
-cronemu_mon(struct tm *wtm, int mon, int mday, int hour, int min)
+waiting4removal_new(job_t j, mach_port_t rp)
{
- if (mon == -1) {
- struct tm workingtm = *wtm;
- int carrytest;
+ struct waiting_for_removal *w4r;
- while (!cronemu_mday(&workingtm, mday, hour, min)) {
- workingtm.tm_mon++;
- workingtm.tm_mday = 1;
- workingtm.tm_hour = 0;
- workingtm.tm_min = 0;
- carrytest = workingtm.tm_mon;
- mktime(&workingtm);
- if (carrytest != workingtm.tm_mon)
- return false;
- }
- *wtm = workingtm;
- return true;
+ if (!job_assumes(j, (w4r = malloc(sizeof(struct waiting_for_removal))) != NULL)) {
+ return false;
}
- if (mon < wtm->tm_mon)
- return false;
+ w4r->reply_port = rp;
- if (mon > wtm->tm_mon) {
- wtm->tm_mon = mon;
- wtm->tm_mday = 1;
- wtm->tm_hour = 0;
- wtm->tm_min = 0;
- }
+ SLIST_INSERT_HEAD(&j->removal_watchers, w4r, sle);
- return cronemu_mday(wtm, mday, hour, min);
+ return true;
}
-bool
-cronemu_mday(struct tm *wtm, int mday, int hour, int min)
+void
+waiting4removal_delete(job_t j, struct waiting_for_removal *w4r)
{
- if (mday == -1) {
- struct tm workingtm = *wtm;
- int carrytest;
-
- while (!cronemu_hour(&workingtm, hour, min)) {
- workingtm.tm_mday++;
- workingtm.tm_hour = 0;
- workingtm.tm_min = 0;
- carrytest = workingtm.tm_mday;
- mktime(&workingtm);
- if (carrytest != workingtm.tm_mday)
- return false;
- }
- *wtm = workingtm;
- return true;
- }
+ (void)job_assumes(j, job_mig_send_signal_reply(w4r->reply_port, 0) == 0);
- if (mday < wtm->tm_mday)
- return false;
-
- if (mday > wtm->tm_mday) {
- wtm->tm_mday = mday;
- wtm->tm_hour = 0;
- wtm->tm_min = 0;
- }
+ SLIST_REMOVE(&j->removal_watchers, w4r, waiting_for_removal, sle);
- return cronemu_hour(wtm, hour, min);
+ free(w4r);
}
-bool
-cronemu_hour(struct tm *wtm, int hour, int min)
+size_t
+get_kern_max_proc(void)
{
- if (hour == -1) {
- struct tm workingtm = *wtm;
- int carrytest;
+ int mib[] = { CTL_KERN, KERN_MAXPROC };
+ int max = 100;
+ size_t max_sz = sizeof(max);
+
+ (void)launchd_assumes(sysctl(mib, 2, &max, &max_sz, NULL, 0) != -1);
+
+ return max;
+}
- while (!cronemu_min(&workingtm, min)) {
- workingtm.tm_hour++;
- workingtm.tm_min = 0;
- carrytest = workingtm.tm_hour;
- mktime(&workingtm);
- if (carrytest != workingtm.tm_hour)
- return false;
+/* See rdar://problem/6271234 */
+void
+eliminate_double_reboot(void)
+{
+ if (unlikely(!pid1_magic)) {
+ return;
+ }
+
+ struct stat sb;
+ const char *argv[] = { _PATH_BSHELL, "/etc/rc.deferred_install", NULL };
+ char *try_again = "Will try again at next boot.";
+ int result = ~0;
+
+ if (unlikely(stat(argv[1], &sb) != -1)) {
+ jobmgr_log(root_jobmgr, LOG_DEBUG | LOG_CONSOLE, "Going to run deferred install script.");
+
+ int wstatus;
+ pid_t p;
+
+ (void)jobmgr_assumes(root_jobmgr, (errno = posix_spawnp(&p, argv[0], NULL, NULL, (char **)argv, environ)) == 0);
+
+ if (errno) {
+ jobmgr_log(root_jobmgr, LOG_WARNING | LOG_CONSOLE, "Couldn't run deferred install script! %s", try_again);
+ goto out;
+ }
+
+ if (!jobmgr_assumes(root_jobmgr, waitpid(p, &wstatus, 0) != -1)) {
+ jobmgr_log(root_jobmgr, LOG_WARNING | LOG_CONSOLE, "Couldn't confirm that deferred install script exited successfully! %s", try_again);
+ goto out;
+ }
+
+ if (jobmgr_assumes(root_jobmgr, WIFEXITED(wstatus) != 0)) {
+ if (jobmgr_assumes(root_jobmgr, (result = WEXITSTATUS(wstatus)) == EXIT_SUCCESS)) {
+ jobmgr_log(root_jobmgr, LOG_DEBUG | LOG_CONSOLE, "Deferred install script completed successfully.");
+ } else {
+ jobmgr_log(root_jobmgr, LOG_WARNING | LOG_CONSOLE, "Deferred install script exited with status %d. %s", WEXITSTATUS(wstatus), try_again);
+ }
+ } else {
+ jobmgr_log(root_jobmgr, LOG_WARNING | LOG_CONSOLE, "Confirmed that deferred install script exited, but couldn't confirm that it was successful. %s", try_again);
}
- *wtm = workingtm;
- return true;
}
+out:
+ if (result == 0) {
+ /* If the unlink(2) was to fail, it would be most likely fail with EBUSY. All the other
+ * failure cases for unlink(2) don't apply when we're running under PID 1 and have verified
+ * that the file exists. Outside of someone deliberately messing with us (like if /etc/rc.deferredinstall
+ * is actually a looping sym-link or a mount point for a filesystem) and I/O errors, we should be good.
+ */
+ if (!jobmgr_assumes(root_jobmgr, unlink(argv[1]) != -1)) {
+ jobmgr_log(root_jobmgr, LOG_WARNING | LOG_CONSOLE, "Deferred install script couldn't be removed!");
+ }
+ }
+}
- if (hour < wtm->tm_hour)
- return false;
-
- if (hour > wtm->tm_hour) {
- wtm->tm_hour = hour;
- wtm->tm_min = 0;
+void
+jetsam_property_setup(launch_data_t obj, const char *key, job_t j)
+{
+ job_log(j, LOG_DEBUG, "Setting Jetsam properties for job...");
+ if (strcasecmp(key, LAUNCH_JOBKEY_JETSAMPRIORITY) == 0 && launch_data_get_type(obj) == LAUNCH_DATA_INTEGER) {
+ j->jetsam_priority = (typeof(j->jetsam_priority))launch_data_get_integer(obj);
+ job_log(j, LOG_DEBUG, "Priority: %d", j->jetsam_priority);
+ } else if (strcasecmp(key, LAUNCH_JOBKEY_JETSAMMEMORYLIMIT) == 0 && launch_data_get_type(obj) == LAUNCH_DATA_INTEGER) {
+ j->jetsam_memlimit = (typeof(j->jetsam_memlimit))launch_data_get_integer(obj);
+ job_log(j, LOG_DEBUG, "Memory limit: %d", j->jetsam_memlimit);
+ } else if (strcasecmp(key, LAUNCH_KEY_JETSAMFRONTMOST) == 0) {
+ /* Ignore. We only recognize this key so we don't complain when we get SpringBoard's request.
+ * You can't set this in a plist.
+ */
+ } else if (strcasecmp(key, LAUNCH_KEY_JETSAMLABEL) == 0) {
+ /* Ignore. This key is present in SpringBoard's request dictionary, so we don't want to
+ * complain about it.
+ */
+ } else {
+ job_log(j, LOG_ERR, "Unknown Jetsam key: %s", key);
+ }
+
+ if (unlikely(!j->jetsam_properties)) {
+ j->jetsam_properties = true;
+ LIST_INSERT_HEAD(&j->mgr->jetsam_jobs, j, jetsam_sle);
+ j->mgr->jetsam_jobs_cnt++;
}
- return cronemu_min(wtm, min);
+ j->jetsam_seq = s_jetsam_sequence_id++;
}
-bool
-cronemu_min(struct tm *wtm, int min)
+int
+launchd_set_jetsam_priorities(launch_data_t priorities)
{
- if (min == -1)
- return true;
-
- if (min < wtm->tm_min)
- return false;
+ if (!launchd_assumes(launch_data_get_type(priorities) == LAUNCH_DATA_ARRAY)) {
+ return EINVAL;
+ }
- if (min > wtm->tm_min) {
- wtm->tm_min = min;
+ jobmgr_t jm = NULL;
+#if !TARGET_OS_EMBEDDED
+ /* For testing. */
+ jm = jobmgr_find_by_name(root_jobmgr, VPROCMGR_SESSION_AQUA);
+ if (!launchd_assumes(jm != NULL)) {
+ return EINVAL;
+ }
+#else
+ /* Since this is for embedded, we can assume that the root job manager holds the Jetsam jobs. */
+ jm = root_jobmgr;
+
+ if (!g_embedded_privileged_action) {
+ return EPERM;
}
+#endif
+
+ size_t npris = launch_data_array_get_count(priorities);
+
+ job_t ji = NULL;
+ size_t i = 0;
+ for (i = 0; i < npris; i++) {
+ launch_data_t ldi = launch_data_array_get_index(priorities, i);
+ if (!launchd_assumes(launch_data_get_type(ldi) == LAUNCH_DATA_DICTIONARY)) {
+ continue;
+ }
+
+ launch_data_t label = NULL;
+ if (!launchd_assumes(label = launch_data_dict_lookup(ldi, LAUNCH_KEY_JETSAMLABEL))) {
+ continue;
+ }
+ const char *_label = launch_data_get_string(label);
+
+ ji = job_find(NULL, _label);
+ if (!launchd_assumes(ji != NULL)) {
+ continue;
+ }
- return true;
+ launch_data_dict_iterate(ldi, (void (*)(launch_data_t, const char *, void *))jetsam_property_setup, ji);
+
+ launch_data_t frontmost = NULL;
+ if ((frontmost = launch_data_dict_lookup(ldi, LAUNCH_KEY_JETSAMFRONTMOST)) && launch_data_get_type(frontmost) == LAUNCH_DATA_BOOL) {
+ ji->jetsam_frontmost = launch_data_get_bool(frontmost);
+ }
+ }
+
+ i = 0;
+ job_t *jobs = (job_t *)calloc(jm->jetsam_jobs_cnt, sizeof(job_t));
+ if (launchd_assumes(jobs != NULL)) {
+ LIST_FOREACH(ji, &jm->jetsam_jobs, jetsam_sle) {
+ if (ji->p) {
+ jobs[i] = ji;
+ i++;
+ }
+ }
+ }
+
+ size_t totalpris = i;
+
+ int result = EINVAL;
+
+ /* It is conceivable that there could be no Jetsam jobs running. */
+ if (totalpris > 0) {
+ /* Yay blocks! */
+ qsort_b((void *)jobs, totalpris, sizeof(job_t), ^ int (const void *lhs, const void *rhs) {
+ job_t _lhs = *(job_t *)lhs;
+ job_t _rhs = *(job_t *)rhs;
+ /* Sort in descending order. (Priority correlates to the soonishness with which you will be killed.) */
+ if (_lhs->jetsam_priority > _rhs->jetsam_priority) {
+ return -1;
+ } else if (_lhs->jetsam_priority < _rhs->jetsam_priority) {
+ return 1;
+ }
+ /* Priority is equal, so sort by sequence ID to maintain LRU order */
+ if( (int)(_lhs->jetsam_seq - _rhs->jetsam_seq) > 0 ) {
+ return 1;
+ } else if( (int)(_lhs->jetsam_seq - _rhs->jetsam_seq) < 0 ) {
+ return -1;
+ }
+
+ return 0;
+ });
+
+ jetsam_priority_entry_t *jpris = (jetsam_priority_entry_t *)calloc(totalpris, sizeof(jetsam_priority_entry_t));
+ if (!launchd_assumes(jpris != NULL)) {
+ result = ENOMEM;
+ } else {
+ for (i = 0; i < totalpris; i++) {
+ jpris[i].pid = jobs[i]->p; /* Subject to time-of-use vs. time-of-check, obviously. */
+ jpris[i].flags |= jobs[i]->jetsam_frontmost ? kJetsamFlagsFrontmost : 0;
+ jpris[i].hiwat_pages = jobs[i]->jetsam_memlimit;
+ }
+
+ (void)launchd_assumes((result = sysctlbyname("kern.memorystatus_priority_list", NULL, NULL, &jpris[0], totalpris * sizeof(jetsam_priority_entry_t))) != -1);
+ result = result != 0 ? errno : 0;
+
+ free(jpris);
+ }
+ }
+
+ if (jobs) {
+ free(jobs);
+ }
+
+ return result;
}