* Copyright (c) 2000-2008 Apple Inc. All rights reserved.
*
* @APPLE_OSREFERENCE_LICENSE_HEADER_START@
- *
+ *
* This file contains Original Code and/or Modifications of Original Code
* as defined in and that are subject to the Apple Public Source License
* Version 2.0 (the 'License'). You may not use this file except in
* unlawful or unlicensed copies of an Apple operating system, or to
* circumvent, violate, or enable the circumvention or violation of, any
* terms of an Apple operating system software license agreement.
- *
+ *
* Please obtain a copy of the License at
* http://www.opensource.apple.com/apsl/ and read it before using this file.
- *
+ *
* The Original Code and all software distributed under the License are
* distributed on an 'AS IS' basis, WITHOUT WARRANTY OF ANY KIND, EITHER
* EXPRESS OR IMPLIED, AND APPLE HEREBY DISCLAIMS ALL SUCH WARRANTIES,
* FITNESS FOR A PARTICULAR PURPOSE, QUIET ENJOYMENT OR NON-INFRINGEMENT.
* Please see the License for the specific language governing rights and
* limitations under the License.
- *
+ *
* @APPLE_OSREFERENCE_LICENSE_HEADER_END@
*/
/* Copyright (c) 1995 NeXT Computer, Inc. All Rights Reserved */
#endif
#include <IOKit/IOBSD.h>
#include <sys/time.h>
+#include <kern/remote_time.h>
-#define HZ 100 /* XXX */
+#define HZ 100 /* XXX */
/* simple lock used to access timezone, tz structure */
lck_spin_t * tz_slock;
lck_grp_t * tz_slock_grp;
lck_attr_t * tz_slock_attr;
-lck_grp_attr_t *tz_slock_grp_attr;
+lck_grp_attr_t *tz_slock_grp_attr;
-static void setthetime(
- struct timeval *tv);
+static void setthetime(
+ struct timeval *tv);
void time_zone_slock_init(void);
+static boolean_t timeval_fixusec(struct timeval *t1);
/*
* Time of day and interval timer support.
/* ARGSUSED */
int
gettimeofday(
- struct proc *p,
- struct gettimeofday_args *uap,
- __unused int32_t *retval)
+ struct proc *p,
+ struct gettimeofday_args *uap,
+ __unused int32_t *retval)
{
int error = 0;
struct timezone ltz; /* local copy */
/* Check that this task is entitled to set the time or it is root */
if (!IOTaskHasEntitlement(current_task(), SETTIME_ENTITLEMENT)) {
-
#if CONFIG_MACF
error = mac_system_check_settime(kauth_cred_get());
- if (error)
- return (error);
+ if (error) {
+ return error;
+ }
#endif
#ifndef CONFIG_EMBEDDED
- if ((error = suser(kauth_cred_get(), &p->p_acflag)))
- return (error);
+ if ((error = suser(kauth_cred_get(), &p->p_acflag))) {
+ return error;
+ }
#endif
}
atv.tv_sec = user_atv.tv_sec;
atv.tv_usec = user_atv.tv_usec;
}
- if (error)
- return (error);
+ if (error) {
+ return error;
+ }
+ }
+ if (uap->tzp && (error = copyin(uap->tzp, (caddr_t)&atz, sizeof(atz)))) {
+ return error;
}
- if (uap->tzp && (error = copyin(uap->tzp, (caddr_t)&atz, sizeof(atz))))
- return (error);
if (uap->tv) {
- timevalfix(&atv);
- if (atv.tv_sec < 0 || (atv.tv_sec == 0 && atv.tv_usec < 0))
- return (EPERM);
+ /* only positive values of sec/usec are accepted */
+ if (atv.tv_sec < 0 || atv.tv_usec < 0) {
+ return EPERM;
+ }
+ if (!timeval_fixusec(&atv)) {
+ return EPERM;
+ }
setthetime(&atv);
}
if (uap->tzp) {
tz = atz;
lck_spin_unlock(tz_slock);
}
- return (0);
+ return 0;
}
static void
setthetime(
- struct timeval *tv)
+ struct timeval *tv)
{
clock_set_calendar_microtime(tv->tv_sec, tv->tv_usec);
}
*/
void
inittodr(
- __unused time_t base)
+ __unused time_t base)
{
- struct timeval tv;
+ struct timeval tv;
/*
* Assertion:
microtime(&tv);
if (tv.tv_sec < 0 || tv.tv_usec < 0) {
- printf ("WARNING: preposterous time in Real Time Clock");
- tv.tv_sec = 0; /* the UNIX epoch */
+ printf("WARNING: preposterous time in Real Time Clock");
+ tv.tv_sec = 0; /* the UNIX epoch */
tv.tv_usec = 0;
setthetime(&tv);
printf(" -- CHECK AND RESET THE DATE!\n");
time_t
boottime_sec(void)
{
- clock_sec_t secs;
- clock_nsec_t nanosecs;
+ clock_sec_t secs;
+ clock_nsec_t nanosecs;
clock_get_boottime_nanotime(&secs, &nanosecs);
- return (secs);
+ return secs;
}
void
boottime_timeval(struct timeval *tv)
{
- clock_sec_t secs;
- clock_usec_t microsecs;
+ clock_sec_t secs;
+ clock_usec_t microsecs;
clock_get_boottime_microtime(&secs, µsecs);
{
struct itimerval aitv;
- if (uap->which > ITIMER_PROF)
- return(EINVAL);
+ if (uap->which > ITIMER_PROF) {
+ return EINVAL;
+ }
bzero(&aitv, sizeof(aitv));
proc_spinlock(p);
switch (uap->which) {
-
case ITIMER_REAL:
/*
* If time for real time timer has passed return 0,
*/
aitv = p->p_realtimer;
if (timerisset(&p->p_rtime)) {
- struct timeval now;
+ struct timeval now;
microuptime(&now);
- if (timercmp(&p->p_rtime, &now, <))
+ if (timercmp(&p->p_rtime, &now, <)) {
timerclear(&aitv.it_value);
- else {
+ } else {
aitv.it_value = p->p_rtime;
timevalsub(&aitv.it_value, &now);
}
- }
- else
+ } else {
timerclear(&aitv.it_value);
+ }
break;
case ITIMER_VIRTUAL:
if (IS_64BIT_PROCESS(p)) {
struct user64_itimerval user_itv;
- bzero(&user_itv, sizeof (user_itv));
+ bzero(&user_itv, sizeof(user_itv));
user_itv.it_interval.tv_sec = aitv.it_interval.tv_sec;
user_itv.it_interval.tv_usec = aitv.it_interval.tv_usec;
user_itv.it_value.tv_sec = aitv.it_value.tv_sec;
user_itv.it_value.tv_usec = aitv.it_value.tv_usec;
- return (copyout((caddr_t)&user_itv, uap->itv, sizeof (user_itv)));
+ return copyout((caddr_t)&user_itv, uap->itv, sizeof(user_itv));
} else {
struct user32_itimerval user_itv;
- bzero(&user_itv, sizeof (user_itv));
+ bzero(&user_itv, sizeof(user_itv));
user_itv.it_interval.tv_sec = aitv.it_interval.tv_sec;
user_itv.it_interval.tv_usec = aitv.it_interval.tv_usec;
user_itv.it_value.tv_sec = aitv.it_value.tv_sec;
user_itv.it_value.tv_usec = aitv.it_value.tv_usec;
- return (copyout((caddr_t)&user_itv, uap->itv, sizeof (user_itv)));
+ return copyout((caddr_t)&user_itv, uap->itv, sizeof(user_itv));
}
}
bzero(&aitv, sizeof(aitv));
- if (uap->which > ITIMER_PROF)
- return (EINVAL);
+ if (uap->which > ITIMER_PROF) {
+ return EINVAL;
+ }
if ((itvp = uap->itv)) {
if (IS_64BIT_PROCESS(p)) {
struct user64_itimerval user_itv;
- if ((error = copyin(itvp, (caddr_t)&user_itv, sizeof (user_itv))))
- return (error);
+ if ((error = copyin(itvp, (caddr_t)&user_itv, sizeof(user_itv)))) {
+ return error;
+ }
aitv.it_interval.tv_sec = user_itv.it_interval.tv_sec;
aitv.it_interval.tv_usec = user_itv.it_interval.tv_usec;
aitv.it_value.tv_sec = user_itv.it_value.tv_sec;
aitv.it_value.tv_usec = user_itv.it_value.tv_usec;
- } else {
+ } else {
struct user32_itimerval user_itv;
- if ((error = copyin(itvp, (caddr_t)&user_itv, sizeof (user_itv))))
- return (error);
+ if ((error = copyin(itvp, (caddr_t)&user_itv, sizeof(user_itv)))) {
+ return error;
+ }
aitv.it_interval.tv_sec = user_itv.it_interval.tv_sec;
aitv.it_interval.tv_usec = user_itv.it_interval.tv_usec;
aitv.it_value.tv_sec = user_itv.it_value.tv_sec;
aitv.it_value.tv_usec = user_itv.it_value.tv_usec;
}
}
- if ((uap->itv = uap->oitv) && (error = getitimer(p, (struct getitimer_args *)uap, retval)))
- return (error);
- if (itvp == 0)
- return (0);
- if (itimerfix(&aitv.it_value) || itimerfix(&aitv.it_interval))
- return (EINVAL);
+ if ((uap->itv = uap->oitv) && (error = getitimer(p, (struct getitimer_args *)uap, retval))) {
+ return error;
+ }
+ if (itvp == 0) {
+ return 0;
+ }
+ if (itimerfix(&aitv.it_value) || itimerfix(&aitv.it_interval)) {
+ return EINVAL;
+ }
switch (uap->which) {
-
case ITIMER_REAL:
proc_spinlock(p);
if (timerisset(&aitv.it_value)) {
timevaladd(&p->p_rtime, &aitv.it_value);
p->p_realtimer = aitv;
if (!thread_call_enter_delayed_with_leeway(p->p_rcall, NULL,
- tvtoabstime(&p->p_rtime), 0, THREAD_CALL_DELAY_USER_NORMAL))
+ tvtoabstime(&p->p_rtime), 0, THREAD_CALL_DELAY_USER_NORMAL)) {
p->p_ractive++;
- } else {
+ }
+ } else {
timerclear(&p->p_rtime);
p->p_realtimer = aitv;
- if (thread_call_cancel(p->p_rcall))
+ if (thread_call_cancel(p->p_rcall)) {
p->p_ractive--;
+ }
}
proc_spinunlock(p);
case ITIMER_VIRTUAL:
- if (timerisset(&aitv.it_value))
+ if (timerisset(&aitv.it_value)) {
task_vtimer_set(p->task, TASK_VTIMER_USER);
- else
+ } else {
task_vtimer_clear(p->task, TASK_VTIMER_USER);
+ }
proc_spinlock(p);
p->p_vtimer_user = aitv;
break;
case ITIMER_PROF:
- if (timerisset(&aitv.it_value))
+ if (timerisset(&aitv.it_value)) {
task_vtimer_set(p->task, TASK_VTIMER_PROF);
- else
+ } else {
task_vtimer_clear(p->task, TASK_VTIMER_PROF);
+ }
proc_spinlock(p);
p->p_vtimer_prof = aitv;
break;
}
- return (0);
+ return 0;
}
/*
*/
proc_spinunlock(p);
- if (r != NULL)
+ if (r != NULL) {
proc_rele(r);
+ }
return;
}
if ((p->p_rtime.tv_sec + 2) >= t.tv_sec) {
for (;;) {
timevaladd(&p->p_rtime, &p->p_realtimer.it_interval);
- if (timercmp(&p->p_rtime, &t, >))
+ if (timercmp(&p->p_rtime, &t, >)) {
break;
+ }
}
} else {
p->p_rtime = p->p_realtimer.it_interval;
assert(p->p_rcall != NULL);
if (!thread_call_enter_delayed_with_leeway(p->p_rcall, NULL, tvtoabstime(&p->p_rtime), 0,
- THREAD_CALL_DELAY_USER_NORMAL)) {
+ THREAD_CALL_DELAY_USER_NORMAL)) {
p->p_ractive++;
}
itimerfix(
struct timeval *tv)
{
-
if (tv->tv_sec < 0 || tv->tv_sec > 100000000 ||
- tv->tv_usec < 0 || tv->tv_usec >= 1000000)
- return (EINVAL);
- return (0);
+ tv->tv_usec < 0 || tv->tv_usec >= 1000000) {
+ return EINVAL;
+ }
+ return 0;
}
int
/* The INT32_MAX limit ensures the timespec is safe for clock_*() functions
* which accept 32-bit ints. */
if (ts->tv_sec < 0 || ts->tv_sec > INT32_MAX ||
- ts->tv_nsec < 0 || (unsigned long long)ts->tv_nsec > NSEC_PER_SEC) {
+ ts->tv_nsec < 0 || (unsigned long long)ts->tv_nsec > NSEC_PER_SEC) {
return 0;
}
return 1;
*/
int
itimerdecr(proc_t p,
- struct itimerval *itp, int usec)
+ struct itimerval *itp, int usec)
{
-
proc_spinlock(p);
-
+
if (itp->it_value.tv_usec < usec) {
if (itp->it_value.tv_sec == 0) {
/* expired, and already in next interval */
usec = 0;
if (timerisset(&itp->it_value)) {
proc_spinunlock(p);
- return (1);
+ return 1;
}
/* expired, exactly at end of interval */
expire:
if (timerisset(&itp->it_interval)) {
itp->it_value = itp->it_interval;
if (itp->it_value.tv_sec > 0) {
- itp->it_value.tv_usec -= usec;
- if (itp->it_value.tv_usec < 0) {
- itp->it_value.tv_usec += 1000000;
- itp->it_value.tv_sec--;
+ itp->it_value.tv_usec -= usec;
+ if (itp->it_value.tv_usec < 0) {
+ itp->it_value.tv_usec += 1000000;
+ itp->it_value.tv_sec--;
}
}
- } else
- itp->it_value.tv_usec = 0; /* sec is already 0 */
+ } else {
+ itp->it_value.tv_usec = 0; /* sec is already 0 */
+ }
proc_spinunlock(p);
- return (0);
+ return 0;
}
/*
struct timeval *t1,
struct timeval *t2)
{
-
t1->tv_sec += t2->tv_sec;
t1->tv_usec += t2->tv_usec;
timevalfix(t1);
struct timeval *t1,
struct timeval *t2)
{
-
t1->tv_sec -= t2->tv_sec;
t1->tv_usec -= t2->tv_usec;
timevalfix(t1);
timevalfix(
struct timeval *t1)
{
-
if (t1->tv_usec < 0) {
t1->tv_sec--;
t1->tv_usec += 1000000;
}
}
+static boolean_t
+timeval_fixusec(
+ struct timeval *t1)
+{
+ assert(t1->tv_usec >= 0);
+ assert(t1->tv_sec >= 0);
+
+ if (t1->tv_usec >= 1000000) {
+ if (os_add_overflow(t1->tv_sec, t1->tv_usec / 1000000, &t1->tv_sec)) {
+ return FALSE;
+ }
+ t1->tv_usec = t1->tv_usec % 1000000;
+ }
+
+ return TRUE;
+}
+
/*
* Return the best possible estimate of the time in the timeval
* to which tvp points.
*/
void
microtime(
- struct timeval *tvp)
+ struct timeval *tvp)
{
- clock_sec_t tv_sec;
- clock_usec_t tv_usec;
+ clock_sec_t tv_sec;
+ clock_usec_t tv_usec;
clock_get_calendar_microtime(&tv_sec, &tv_usec);
void
microtime_with_abstime(
- struct timeval *tvp, uint64_t *abstime)
+ struct timeval *tvp, uint64_t *abstime)
{
- clock_sec_t tv_sec;
- clock_usec_t tv_usec;
+ clock_sec_t tv_sec;
+ clock_usec_t tv_usec;
clock_get_calendar_absolute_and_microtime(&tv_sec, &tv_usec, abstime);
void
microuptime(
- struct timeval *tvp)
+ struct timeval *tvp)
{
- clock_sec_t tv_sec;
- clock_usec_t tv_usec;
+ clock_sec_t tv_sec;
+ clock_usec_t tv_usec;
clock_get_system_microtime(&tv_sec, &tv_usec);
nanotime(
struct timespec *tsp)
{
- clock_sec_t tv_sec;
- clock_nsec_t tv_nsec;
+ clock_sec_t tv_sec;
+ clock_nsec_t tv_nsec;
clock_get_calendar_nanotime(&tv_sec, &tv_nsec);
nanouptime(
struct timespec *tsp)
{
- clock_sec_t tv_sec;
- clock_nsec_t tv_nsec;
+ clock_sec_t tv_sec;
+ clock_nsec_t tv_nsec;
clock_get_system_nanotime(&tv_sec, &tv_nsec);
uint64_t
tvtoabstime(
- struct timeval *tvp)
+ struct timeval *tvp)
{
- uint64_t result, usresult;
+ uint64_t result, usresult;
clock_interval_to_absolutetime_interval(
- tvp->tv_sec, NSEC_PER_SEC, &result);
+ tvp->tv_sec, NSEC_PER_SEC, &result);
clock_interval_to_absolutetime_interval(
- tvp->tv_usec, NSEC_PER_USEC, &usresult);
+ tvp->tv_usec, NSEC_PER_USEC, &usresult);
- return (result + usresult);
+ return result + usresult;
}
uint64_t
rv = 1;
}
- return (rv);
+ return rv;
}
/*
*lasttime = tv;
*curpps = 0;
rv = 1;
- } else if (maxpps < 0)
+ } else if (maxpps < 0) {
rv = 1;
- else if (*curpps < maxpps)
+ } else if (*curpps < maxpps) {
rv = 1;
- else
+ } else {
rv = 0;
+ }
#if 1 /* DIAGNOSTIC? */
/* be careful about wrap-around */
- if (*curpps + 1 > 0)
+ if (*curpps + 1 > 0) {
*curpps = *curpps + 1;
+ }
#else
/*
* assume that there's not too many calls to this function.
*curpps = *curpps + 1;
#endif
- return (rv);
+ return rv;
}
#endif /* NETWORKING */
/* Allocate the spin lock */
tz_slock = lck_spin_alloc_init(tz_slock_grp, tz_slock_attr);
}
+
+int
+__mach_bridge_remote_time(__unused struct proc *p, struct __mach_bridge_remote_time_args *mbrt_args, uint64_t *retval)
+{
+ *retval = mach_bridge_remote_time(mbrt_args->local_timestamp);
+ return 0;
+}