]> git.saurik.com Git - apple/xnu.git/blobdiff - osfmk/kern/clock.c
xnu-1504.15.3.tar.gz
[apple/xnu.git] / osfmk / kern / clock.c
index 4085c5883b7670b5d26f9cb39716a0a2f379c6a0..fd2e29797517ebd34a5d246cdc3f7bd09c8e04e6 100644 (file)
@@ -1,23 +1,29 @@
 /*
- * Copyright (c) 2000-2005 Apple Computer, Inc. All rights reserved.
+ * Copyright (c) 2000-2008 Apple Inc. All rights reserved.
  *
- * @APPLE_LICENSE_HEADER_START@
+ * @APPLE_OSREFERENCE_LICENSE_HEADER_START@
  * 
- * The contents of this file constitute Original Code as defined in and
- * are subject to the Apple Public Source License Version 1.1 (the
- * "License").  You may not use this file except in compliance with the
- * License.  Please obtain a copy of the License at
- * http://www.apple.com/publicsource and read it before using this file.
+ * This file contains Original Code and/or Modifications of Original Code
+ * as defined in and that are subject to the Apple Public Source License
+ * Version 2.0 (the 'License'). You may not use this file except in
+ * compliance with the License. The rights granted to you under the License
+ * may not be used to create, or enable the creation or redistribution of,
+ * unlawful or unlicensed copies of an Apple operating system, or to
+ * circumvent, violate, or enable the circumvention or violation of, any
+ * terms of an Apple operating system software license agreement.
  * 
- * This Original Code and all software distributed under the License are
- * distributed on an "AS IS" basis, WITHOUT WARRANTY OF ANY KIND, EITHER
+ * Please obtain a copy of the License at
+ * http://www.opensource.apple.com/apsl/ and read it before using this file.
+ * 
+ * The Original Code and all software distributed under the License are
+ * distributed on an 'AS IS' basis, WITHOUT WARRANTY OF ANY KIND, EITHER
  * EXPRESS OR IMPLIED, AND APPLE HEREBY DISCLAIMS ALL SUCH WARRANTIES,
  * INCLUDING WITHOUT LIMITATION, ANY WARRANTIES OF MERCHANTABILITY,
- * FITNESS FOR A PARTICULAR PURPOSE OR NON-INFRINGEMENT.  Please see the
- * License for the specific language governing rights and limitations
- * under the License.
+ * FITNESS FOR A PARTICULAR PURPOSE, QUIET ENJOYMENT OR NON-INFRINGEMENT.
+ * Please see the License for the specific language governing rights and
+ * limitations under the License.
  * 
- * @APPLE_LICENSE_HEADER_END@
+ * @APPLE_OSREFERENCE_LICENSE_HEADER_END@
  */
 /*
  * @OSF_COPYRIGHT@
 #include <mach/mach_traps.h>
 #include <mach/mach_time.h>
 
+uint32_t       hz_tick_interval = 1;
+
+
 decl_simple_lock_data(static,clock_lock)
 
+#define clock_lock()   \
+       simple_lock(&clock_lock)
+
+#define clock_unlock() \
+       simple_unlock(&clock_lock)
+
+#define clock_lock_init()      \
+       simple_lock_init(&clock_lock, 0)
+
+
 /*
  *     Time of day (calendar) variables.
  *
@@ -53,9 +72,35 @@ decl_simple_lock_data(static,clock_lock)
  *     where CONV converts absolute time units into seconds and a fraction.
  */
 static struct clock_calend {
-       uint64_t                        epoch;
-       uint64_t                        offset;
-}                                      clock_calend;
+
+       uint64_t        epoch;
+       uint64_t        offset;
+
+       int32_t         adjdelta;       /* Nanosecond time delta for this adjustment period */
+       uint64_t        adjstart;       /* Absolute time value for start of this adjustment period */
+       uint32_t        adjoffset;      /* Absolute time offset for this adjustment period as absolute value */
+} clock_calend;
+
+#if    CONFIG_DTRACE
+
+/*
+ *     Unlocked calendar flipflop; this is used to track a clock_calend such
+ *     that we can safely access a snapshot of a valid  clock_calend structure
+ *     without needing to take any locks to do it.
+ *
+ *     The trick is to use a generation count and set the low bit when it is
+ *     being updated/read; by doing this, we guarantee, through use of the
+ *     hw_atomic functions, that the generation is incremented when the bit
+ *     is cleared atomically (by using a 1 bit add).
+ */
+static struct unlocked_clock_calend {
+       struct clock_calend     calend;         /* copy of calendar */
+       uint32_t                gen;            /* generation count */
+} flipflop[ 2];
+
+static void clock_track_calend_nowait(void);
+
+#endif
 
 /*
  *     Calendar adjustment variables and values.
@@ -64,21 +109,16 @@ static struct clock_calend {
 #define calend_adjskew         (40 * NSEC_PER_USEC)    /* "standard" skew, ns / period */
 #define        calend_adjbig           (NSEC_PER_SEC)                  /* use 10x skew above adjbig ns */
 
-static uint64_t                        calend_adjstart;                /* Absolute time value for start of this adjustment period */
-static uint32_t                        calend_adjoffset;               /* Absolute time offset for this adjustment period as absolute value */
-
-static int32_t                 calend_adjdelta;                /* Nanosecond time delta for this adjustment period */
-static int64_t                 calend_adjtotal;                /* Nanosecond remaining total adjustment */
-
-static uint64_t                        calend_adjdeadline;             /* Absolute time value for next adjustment period */
-static uint32_t                        calend_adjinterval;             /* Absolute time interval of adjustment period */
+static int64_t                         calend_adjtotal;                /* Nanosecond remaining total adjustment */
+static uint64_t                                calend_adjdeadline;             /* Absolute time value for next adjustment period */
+static uint32_t                                calend_adjinterval;             /* Absolute time interval of adjustment period */
 
 static timer_call_data_t       calend_adjcall;
 static uint32_t                                calend_adjactive;
 
 static uint32_t                calend_set_adjustment(
-                                               int32_t                 *secs,
-                                               int32_t                 *microsecs);
+                                               long                    *secs,
+                                               int                             *microsecs);
 
 static void                    calend_adjust_call(void);
 static uint32_t                calend_adjust(void);
@@ -100,7 +140,7 @@ MACRO_END
 
 #define TIME_SUB(rsecs, secs, rfrac, frac, unit)       \
 MACRO_BEGIN                                                                                    \
-       if ((int32_t)((rfrac) -= (frac)) < 0) {                 \
+       if ((int)((rfrac) -= (frac)) < 0) {                             \
                (rfrac) += (unit);                                                      \
                (rsecs) -= 1;                                                           \
        }                                                                                               \
@@ -115,7 +155,7 @@ MACRO_END
 void
 clock_config(void)
 {
-       simple_lock_init(&clock_lock, 0);
+       clock_lock_init();
 
        timer_call_setup(&calend_adjcall, (timer_call_func_t)calend_adjust_call, NULL);
        thread_call_setup(&calend_wakecall, (thread_call_func_t)IOKitResetTime, NULL);
@@ -153,7 +193,10 @@ clock_timebase_init(void)
        uint64_t        abstime;
 
        nanoseconds_to_absolutetime(calend_adjperiod, &abstime);
-       calend_adjinterval = abstime;
+       calend_adjinterval = (uint32_t)abstime;
+
+       nanoseconds_to_absolutetime(NSEC_PER_SEC / 100, &abstime);
+       hz_tick_interval = (uint32_t)abstime;
 
        sched_timebase_init();
 }
@@ -189,27 +232,27 @@ mach_timebase_info_trap(
  */
 void
 clock_get_calendar_microtime(
-       uint32_t                        *secs,
-       uint32_t                        *microsecs)
+       clock_sec_t                     *secs,
+       clock_usec_t            *microsecs)
 {
        uint64_t                now;
        spl_t                   s;
 
        s = splclock();
-       simple_lock(&clock_lock);
+       clock_lock();
 
        now = mach_absolute_time();
 
-       if (calend_adjdelta < 0) {
+       if (clock_calend.adjdelta < 0) {
                uint32_t        t32;
 
-               if (now > calend_adjstart) {
-                       t32 = now - calend_adjstart;
+               if (now > clock_calend.adjstart) {
+                       t32 = (uint32_t)(now - clock_calend.adjstart);
 
-                       if (t32 > calend_adjoffset)
-                               now -= calend_adjoffset;
+                       if (t32 > clock_calend.adjoffset)
+                               now -= clock_calend.adjoffset;
                        else
-                               now = calend_adjstart;
+                               now = clock_calend.adjstart;
                }
        }
 
@@ -217,9 +260,9 @@ clock_get_calendar_microtime(
 
        absolutetime_to_microtime(now, secs, microsecs);
 
-       *secs += clock_calend.epoch;
+       *secs += (clock_sec_t)clock_calend.epoch;
 
-       simple_unlock(&clock_lock);
+       clock_unlock();
        splx(s);
 }
 
@@ -235,27 +278,27 @@ clock_get_calendar_microtime(
  */
 void
 clock_get_calendar_nanotime(
-       uint32_t                        *secs,
-       uint32_t                        *nanosecs)
+       clock_sec_t                     *secs,
+       clock_nsec_t            *nanosecs)
 {
        uint64_t                now;
        spl_t                   s;
 
        s = splclock();
-       simple_lock(&clock_lock);
+       clock_lock();
 
        now = mach_absolute_time();
 
-       if (calend_adjdelta < 0) {
+       if (clock_calend.adjdelta < 0) {
                uint32_t        t32;
 
-               if (now > calend_adjstart) {
-                       t32 = now - calend_adjstart;
+               if (now > clock_calend.adjstart) {
+                       t32 = (uint32_t)(now - clock_calend.adjstart);
 
-                       if (t32 > calend_adjoffset)
-                               now -= calend_adjoffset;
+                       if (t32 > clock_calend.adjoffset)
+                               now -= clock_calend.adjoffset;
                        else
-                               now = calend_adjstart;
+                               now = clock_calend.adjstart;
                }
        }
 
@@ -264,9 +307,9 @@ clock_get_calendar_nanotime(
        absolutetime_to_microtime(now, secs, nanosecs);
        *nanosecs *= NSEC_PER_USEC;
 
-       *secs += clock_calend.epoch;
+       *secs += (clock_sec_t)clock_calend.epoch;
 
-       simple_unlock(&clock_lock);
+       clock_unlock();
        splx(s);
 }
 
@@ -283,40 +326,40 @@ clock_get_calendar_nanotime(
  */
 void
 clock_gettimeofday(
-       uint32_t                        *secs,
-       uint32_t                        *microsecs)
+       clock_sec_t             *secs,
+       clock_usec_t    *microsecs)
 {
        uint64_t                now;
        spl_t                   s;
 
        s = splclock();
-       simple_lock(&clock_lock);
+       clock_lock();
 
        now = mach_absolute_time();
 
-       if (calend_adjdelta >= 0) {
+       if (clock_calend.adjdelta >= 0) {
                clock_gettimeofday_set_commpage(now, clock_calend.epoch, clock_calend.offset, secs, microsecs);
        }
        else {
                uint32_t        t32;
 
-               if (now > calend_adjstart) {
-                       t32 = now - calend_adjstart;
+               if (now > clock_calend.adjstart) {
+                       t32 = (uint32_t)(now - clock_calend.adjstart);
 
-                       if (t32 > calend_adjoffset)
-                               now -= calend_adjoffset;
+                       if (t32 > clock_calend.adjoffset)
+                               now -= clock_calend.adjoffset;
                        else
-                               now = calend_adjstart;
+                               now = clock_calend.adjstart;
                }
 
                now += clock_calend.offset;
 
                absolutetime_to_microtime(now, secs, microsecs);
 
-               *secs += clock_calend.epoch;
+               *secs += (clock_sec_t)clock_calend.epoch;
        }
 
-       simple_unlock(&clock_lock);
+       clock_unlock();
        splx(s);
 }
 
@@ -334,20 +377,20 @@ clock_gettimeofday(
  */
 void
 clock_set_calendar_microtime(
-       uint32_t                        secs,
-       uint32_t                        microsecs)
+       clock_sec_t                     secs,
+       clock_usec_t            microsecs)
 {
-       uint32_t                sys, microsys;
-       uint32_t                newsecs;
-       spl_t                   s;
+       clock_sec_t                     sys;
+       clock_usec_t            microsys;
+       clock_sec_t                     newsecs;
+       spl_t                           s;
 
-       newsecs = (microsecs < 500*USEC_PER_SEC)?
-                                               secs: secs + 1;
+       newsecs = (microsecs < 500*USEC_PER_SEC)? secs: secs + 1;
 
        s = splclock();
-       simple_lock(&clock_lock);
+       clock_lock();
 
-    commpage_set_timestamp(0,0,0);
+       commpage_disable_timestamp();
 
        /*
         *      Calculate the new calendar epoch based on
@@ -370,9 +413,9 @@ clock_set_calendar_microtime(
        /*
         *      Cancel any adjustment in progress.
         */
-       calend_adjdelta = calend_adjtotal = 0;
+       calend_adjtotal = clock_calend.adjdelta = 0;
 
-       simple_unlock(&clock_lock);
+       clock_unlock();
 
        /*
         *      Set the new value for the platform clock.
@@ -385,6 +428,10 @@ clock_set_calendar_microtime(
         *      Send host notifications.
         */
        host_notify_calendar_change();
+       
+#if CONFIG_DTRACE
+       clock_track_calend_nowait();
+#endif
 }
 
 /*
@@ -399,16 +446,16 @@ clock_set_calendar_microtime(
 void
 clock_initialize_calendar(void)
 {
-       uint32_t                sys, microsys;
-       uint32_t                microsecs = 0, secs = PEGetGMTTimeOfDay();
-       spl_t                   s;
+       clock_sec_t                     sys, secs = PEGetGMTTimeOfDay();
+       clock_usec_t            microsys, microsecs = 0;
+       spl_t                           s;
 
        s = splclock();
-       simple_lock(&clock_lock);
+       clock_lock();
 
-    commpage_set_timestamp(0,0,0);
+       commpage_disable_timestamp();
 
-       if ((int32_t)secs >= (int32_t)clock_boottime) {
+       if ((long)secs >= (long)clock_boottime) {
                /*
                 *      Initialize the boot time based on the platform clock.
                 */
@@ -431,16 +478,20 @@ clock_initialize_calendar(void)
                /*
                 *       Cancel any adjustment in progress.
                 */
-               calend_adjdelta = calend_adjtotal = 0;
+               calend_adjtotal = clock_calend.adjdelta = 0;
        }
 
-       simple_unlock(&clock_lock);
+       clock_unlock();
        splx(s);
 
        /*
         *      Send host notifications.
         */
        host_notify_calendar_change();
+       
+#if CONFIG_DTRACE
+       clock_track_calend_nowait();
+#endif
 }
 
 /*
@@ -450,11 +501,19 @@ clock_initialize_calendar(void)
  */
 void
 clock_get_boottime_nanotime(
-       uint32_t                        *secs,
-       uint32_t                        *nanosecs)
+       clock_sec_t                     *secs,
+       clock_nsec_t            *nanosecs)
 {
-       *secs = clock_boottime;
+       spl_t   s;
+
+       s = splclock();
+       clock_lock();
+
+       *secs = (clock_sec_t)clock_boottime;
        *nanosecs = 0;
+
+       clock_unlock();
+       splx(s);
 }
 
 /*
@@ -467,14 +526,14 @@ clock_get_boottime_nanotime(
  */
 void
 clock_adjtime(
-       int32_t         *secs,
-       int32_t         *microsecs)
+       long            *secs,
+       int                     *microsecs)
 {
        uint32_t        interval;
        spl_t           s;
 
        s = splclock();
-       simple_lock(&clock_lock);
+       clock_lock();
 
        interval = calend_set_adjustment(secs, microsecs);
        if (interval != 0) {
@@ -486,14 +545,14 @@ clock_adjtime(
        if (timer_call_cancel(&calend_adjcall))
                calend_adjactive--;
 
-       simple_unlock(&clock_lock);
+       clock_unlock();
        splx(s);
 }
 
 static uint32_t
 calend_set_adjustment(
-       int32_t                         *secs,
-       int32_t                         *microsecs)
+       long                    *secs,
+       int                             *microsecs)
 {
        uint64_t                now, t64;
        int64_t                 total, ototal;
@@ -501,7 +560,7 @@ calend_set_adjustment(
 
        total = (int64_t)*secs * NSEC_PER_SEC + *microsecs * NSEC_PER_USEC;
 
-    commpage_set_timestamp(0,0,0);
+       commpage_disable_timestamp();
 
        now = mach_absolute_time();
 
@@ -514,39 +573,43 @@ calend_set_adjustment(
                        if (total > calend_adjbig)
                                delta *= 10;
                        if (delta > total)
-                               delta = total;
+                               delta = (int32_t)total;
 
                        nanoseconds_to_absolutetime((uint64_t)delta, &t64);
-                       calend_adjoffset = t64;
+                       clock_calend.adjoffset = (uint32_t)t64;
                }
                else {
                        if (total < -calend_adjbig)
                                delta *= 10;
                        delta = -delta;
                        if (delta < total)
-                               delta = total;
+                               delta = (int32_t)total;
 
-                       calend_adjstart = now;
+                       clock_calend.adjstart = now;
 
                        nanoseconds_to_absolutetime((uint64_t)-delta, &t64);
-                       calend_adjoffset = t64;
+                       clock_calend.adjoffset = (uint32_t)t64;
                }
 
                calend_adjtotal = total;
-               calend_adjdelta = delta;
+               clock_calend.adjdelta = delta;
 
                interval = calend_adjinterval;
        }
        else
-               calend_adjdelta = calend_adjtotal = 0;
+               calend_adjtotal = clock_calend.adjdelta = 0;
 
        if (ototal != 0) {
-               *secs = ototal / NSEC_PER_SEC;
-               *microsecs = (ototal % NSEC_PER_SEC) / NSEC_PER_USEC;
+               *secs = (long)(ototal / NSEC_PER_SEC);
+               *microsecs = (int)((ototal % NSEC_PER_SEC) / NSEC_PER_USEC);
        }
        else
                *secs = *microsecs = 0;
 
+#if CONFIG_DTRACE
+       clock_track_calend_nowait();
+#endif
+       
        return (interval);
 }
 
@@ -557,20 +620,19 @@ calend_adjust_call(void)
        spl_t           s;
 
        s = splclock();
-       simple_lock(&clock_lock);
+       clock_lock();
 
        if (--calend_adjactive == 0) {
                interval = calend_adjust();
                if (interval != 0) {
-                       clock_deadline_for_periodic_event(interval, mach_absolute_time(),
-                                                                                                                               &calend_adjdeadline);
+                       clock_deadline_for_periodic_event(interval, mach_absolute_time(), &calend_adjdeadline);
 
                        if (!timer_call_enter(&calend_adjcall, calend_adjdeadline))
                                calend_adjactive++;
                }
        }
 
-       simple_unlock(&clock_lock);
+       clock_unlock();
        splx(s);
 }
 
@@ -581,42 +643,46 @@ calend_adjust(void)
        int32_t                 delta;
        uint32_t                interval = 0;
 
-    commpage_set_timestamp(0,0,0);
+       commpage_disable_timestamp();
 
        now = mach_absolute_time();
 
-       delta = calend_adjdelta;
+       delta = clock_calend.adjdelta;
 
        if (delta > 0) {
-               clock_calend.offset += calend_adjoffset;
+               clock_calend.offset += clock_calend.adjoffset;
 
                calend_adjtotal -= delta;
                if (delta > calend_adjtotal) {
-                       calend_adjdelta = delta = calend_adjtotal;
+                       clock_calend.adjdelta = delta = (int32_t)calend_adjtotal;
 
                        nanoseconds_to_absolutetime((uint64_t)delta, &t64);
-                       calend_adjoffset = t64;
+                       clock_calend.adjoffset = (uint32_t)t64;
                }
        }
        else
        if (delta < 0) {
-               clock_calend.offset -= calend_adjoffset;
+               clock_calend.offset -= clock_calend.adjoffset;
 
                calend_adjtotal -= delta;
                if (delta < calend_adjtotal) {
-                       calend_adjdelta = delta = calend_adjtotal;
+                       clock_calend.adjdelta = delta = (int32_t)calend_adjtotal;
 
                        nanoseconds_to_absolutetime((uint64_t)-delta, &t64);
-                       calend_adjoffset = t64;
+                       clock_calend.adjoffset = (uint32_t)t64;
                }
 
-               if (calend_adjdelta != 0)
-                       calend_adjstart = now;
+               if (clock_calend.adjdelta != 0)
+                       clock_calend.adjstart = now;
        }
-
-       if (calend_adjdelta != 0)
+       
+       if (clock_calend.adjdelta != 0)
                interval = calend_adjinterval;
 
+#if CONFIG_DTRACE
+       clock_track_calend_nowait();
+#endif
+
        return (interval);
 }
 
@@ -747,3 +813,102 @@ clock_deadline_for_periodic_event(
                        *deadline = abstime + interval;
        }
 }
+
+#if    CONFIG_DTRACE
+
+/*
+ * clock_get_calendar_nanotime_nowait
+ *
+ * Description:        Non-blocking version of clock_get_calendar_nanotime()
+ *
+ * Notes:      This function operates by separately tracking calendar time
+ *             updates using a two element structure to copy the calendar
+ *             state, which may be asynchronously modified.  It utilizes
+ *             barrier instructions in the tracking process and in the local
+ *             stable snapshot process in order to ensure that a consistent
+ *             snapshot is used to perform the calculation.
+ */
+void
+clock_get_calendar_nanotime_nowait(
+       clock_sec_t                     *secs,
+       clock_nsec_t            *nanosecs)
+{
+       int i = 0;
+       uint64_t                now;
+       struct unlocked_clock_calend stable;
+
+       for (;;) {
+               stable = flipflop[i];           /* take snapshot */
+
+               /*
+                * Use a barrier instructions to ensure atomicity.  We AND
+                * off the "in progress" bit to get the current generation
+                * count.
+                */
+               (void)hw_atomic_and(&stable.gen, ~(uint32_t)1);
+
+               /*
+                * If an update _is_ in progress, the generation count will be
+                * off by one, if it _was_ in progress, it will be off by two,
+                * and if we caught it at a good time, it will be equal (and
+                * our snapshot is threfore stable).
+                */
+               if (flipflop[i].gen == stable.gen)
+                       break;
+
+               /* Switch to the oher element of the flipflop, and try again. */
+               i ^= 1;
+       }
+
+       now = mach_absolute_time();
+
+       if (stable.calend.adjdelta < 0) {
+               uint32_t        t32;
+
+               if (now > stable.calend.adjstart) {
+                       t32 = (uint32_t)(now - stable.calend.adjstart);
+
+                       if (t32 > stable.calend.adjoffset)
+                               now -= stable.calend.adjoffset;
+                       else
+                               now = stable.calend.adjstart;
+               }
+       }
+
+       now += stable.calend.offset;
+
+       absolutetime_to_microtime(now, secs, nanosecs);
+       *nanosecs *= NSEC_PER_USEC;
+
+       *secs += (clock_sec_t)stable.calend.epoch;
+}
+
+static void 
+clock_track_calend_nowait(void)
+{
+       int i;
+
+       for (i = 0; i < 2; i++) {
+               struct clock_calend tmp = clock_calend;
+
+               /*
+                * Set the low bit if the generation count; since we use a
+                * barrier instruction to do this, we are guaranteed that this
+                * will flag an update in progress to an async caller trying
+                * to examine the contents.
+                */
+               (void)hw_atomic_or(&flipflop[i].gen, 1);
+
+               flipflop[i].calend = tmp;
+
+               /*
+                * Increment the generation count to clear the low bit to
+                * signal completion.  If a caller compares the generation
+                * count after taking a copy while in progress, the count
+                * will be off by two.
+                */
+               (void)hw_atomic_add(&flipflop[i].gen, 1);
+       }
+}
+
+#endif /* CONFIG_DTRACE */