--- /dev/null
+/*
+ * Copyright (c) 2012 Apple Inc. All rights reserved.
+ *
+ * @APPLE_OSREFERENCE_LICENSE_HEADER_START@
+ *
+ * This file contains Original Code and/or Modifications of Original Code
+ * as defined in and that are subject to the Apple Public Source License
+ * Version 2.0 (the 'License'). You may not use this file except in
+ * compliance with the License. The rights granted to you under the License
+ * may not be used to create, or enable the creation or redistribution of,
+ * unlawful or unlicensed copies of an Apple operating system, or to
+ * circumvent, violate, or enable the circumvention or violation of, any
+ * terms of an Apple operating system software license agreement.
+ *
+ * Please obtain a copy of the License at
+ * http://www.opensource.apple.com/apsl/ and read it before using this file.
+ *
+ * The Original Code and all software distributed under the License are
+ * distributed on an 'AS IS' basis, WITHOUT WARRANTY OF ANY KIND, EITHER
+ * EXPRESS OR IMPLIED, AND APPLE HEREBY DISCLAIMS ALL SUCH WARRANTIES,
+ * INCLUDING WITHOUT LIMITATION, ANY WARRANTIES OF MERCHANTABILITY,
+ * FITNESS FOR A PARTICULAR PURPOSE, QUIET ENJOYMENT OR NON-INFRINGEMENT.
+ * Please see the License for the specific language governing rights and
+ * limitations under the License.
+ *
+ * @APPLE_OSREFERENCE_LICENSE_HEADER_END@
+ */
+
+#include <mach/mach_types.h>
+#include <kern/processor.h>
+#include <kern/thread.h>
+#include <kern/assert.h>
+#include <kern/locks.h>
+#include <sys/errno.h>
+
+#include <kern/kpc.h>
+
+
+/* global for whether to read PMCs on context switch */
+int kpc_threads_counting;
+
+/* current config and number of counters in that config */
+static uint32_t kpc_thread_classes = 0;
+static uint32_t kpc_thread_classes_count = 0;
+
+static lck_grp_attr_t *kpc_thread_lckgrp_attr = NULL;
+static lck_grp_t *kpc_thread_lckgrp = NULL;
+static lck_mtx_t kpc_thread_lock;
+
+void kpc_thread_init(void);
+
+void
+kpc_thread_init(void)
+{
+ kpc_thread_lckgrp_attr = lck_grp_attr_alloc_init();
+ kpc_thread_lckgrp = lck_grp_alloc_init("kpc", kpc_thread_lckgrp_attr);
+ lck_mtx_init(&kpc_thread_lock, kpc_thread_lckgrp, LCK_ATTR_NULL);
+}
+
+uint32_t
+kpc_get_thread_counting(void)
+{
+ uint32_t kpc_thread_classes_tmp;
+ int kpc_threads_counting_tmp;
+
+ /* Make sure we get a consistent snapshot of these values */
+ lck_mtx_lock(&kpc_thread_lock);
+
+ kpc_thread_classes_tmp = kpc_thread_classes;
+ kpc_threads_counting_tmp = kpc_threads_counting;
+
+ lck_mtx_unlock(&kpc_thread_lock);
+
+ if( kpc_threads_counting_tmp )
+ return kpc_thread_classes_tmp;
+ else
+ return 0;
+}
+
+int
+kpc_set_thread_counting(uint32_t classes)
+{
+ uint32_t count;
+
+ lck_mtx_lock(&kpc_thread_lock);
+
+ count = kpc_get_counter_count(classes);
+
+ if( (classes == 0)
+ || (count == 0) )
+ {
+ /* shut down */
+ kpc_threads_counting = FALSE;
+ }
+ else
+ {
+ /* stash the config */
+ kpc_thread_classes = classes;
+
+ /* work out the size */
+ kpc_thread_classes_count = count;
+ assert(kpc_thread_classes_count <= KPC_MAX_COUNTERS);
+
+ /* enable switch */
+ kpc_threads_counting = TRUE;
+
+ /* and schedule an AST for this thread... */
+ if( !current_thread()->kpc_buf )
+ {
+ current_thread()->t_chud |= T_KPC_ALLOC;
+ act_set_kperf(current_thread());
+ }
+ }
+
+ lck_mtx_unlock(&kpc_thread_lock);
+
+ return 0;
+}
+
+/* snapshot current PMCs and update counters in the current thread */
+static void
+kpc_update_thread_counters( thread_t thread )
+{
+ uint32_t i;
+ uint64_t *tmp = NULL;
+ cpu_data_t *cpu = NULL;
+
+/* TODO: Fix this...*/
+#if defined (__x86_64__)
+ cpu = current_cpu_datap();
+#else
+#error architecture not yet supported
+#endif
+
+ /* 1. stash current PMCs into latest CPU block */
+ kpc_get_cpu_counters( FALSE, kpc_thread_classes,
+ NULL, cpu->cpu_kpc_buf[1] );
+
+ /* 2. apply delta to old thread */
+ if( thread->kpc_buf )
+ for( i = 0; i < kpc_thread_classes_count; i++ )
+ thread->kpc_buf[i] += cpu->cpu_kpc_buf[1][i] - cpu->cpu_kpc_buf[0][i];
+
+
+ /* schedule any necessary allocations */
+ if( !current_thread()->kpc_buf )
+ {
+ current_thread()->t_chud |= T_KPC_ALLOC;
+ act_set_kperf(current_thread());
+ }
+
+ /* 3. switch the PMC block pointers */
+ tmp = cpu->cpu_kpc_buf[1];
+ cpu->cpu_kpc_buf[1] = cpu->cpu_kpc_buf[0];
+ cpu->cpu_kpc_buf[0] = tmp;
+}
+
+void
+kpc_switch_context( thread_t old, thread_t new __unused )
+{
+ kpc_update_thread_counters( old );
+}
+
+/* get counter values for a thread */
+int
+kpc_get_curthread_counters(uint32_t *inoutcount, uint64_t *buf)
+{
+ thread_t thread = current_thread();
+ boolean_t enabled;
+
+ /* buffer too small :( */
+ if( *inoutcount < kpc_thread_classes_count )
+ return EINVAL;
+
+ /* copy data and actual size */
+ if( !thread->kpc_buf )
+ return EINVAL;
+
+ enabled = ml_set_interrupts_enabled(FALSE);
+
+ /* snap latest version of counters for this thread */
+ kpc_update_thread_counters( current_thread() );
+
+ /* copy out */
+ memcpy( buf, thread->kpc_buf,
+ kpc_thread_classes_count * sizeof(*buf) );
+ *inoutcount = kpc_thread_classes_count;
+
+ ml_set_interrupts_enabled(enabled);
+
+ return 0;
+}
+
+
+void
+kpc_thread_create(thread_t thread)
+{
+ /* nothing to do if we're not counting */
+ if(!kpc_threads_counting)
+ return;
+
+ /* give the new thread a counterbuf */
+ thread->kpc_buf = kpc_counterbuf_alloc();
+}
+
+void
+kpc_thread_destroy(thread_t thread)
+{
+ uint64_t *buf = NULL;
+
+ /* usual case: no kpc buf, just return */
+ if( !thread->kpc_buf )
+ return;
+
+ /* otherwise, don't leak */
+ buf = thread->kpc_buf;
+ thread->kpc_buf = NULL;
+ kpc_counterbuf_free(buf);
+}
+
+/* ast callback on a thread */
+void
+kpc_thread_ast_handler( thread_t thread )
+{
+ /* see if we want an alloc */
+ if( thread->t_chud & T_KPC_ALLOC )
+ thread->kpc_buf = kpc_counterbuf_alloc();
+}