]> git.saurik.com Git - apple/xnu.git/blobdiff - osfmk/console/serial_console.c
xnu-3789.1.32.tar.gz
[apple/xnu.git] / osfmk / console / serial_console.c
diff --git a/osfmk/console/serial_console.c b/osfmk/console/serial_console.c
new file mode 100644 (file)
index 0000000..ec13979
--- /dev/null
@@ -0,0 +1,590 @@
+/*
+ * Copyright (c) 2000-2006 Apple Computer, Inc. All rights reserved.
+ *
+ * @APPLE_OSREFERENCE_LICENSE_HEADER_START@
+ *
+ * This file contains Original Code and/or Modifications of Original Code
+ * as defined in and that are subject to the Apple Public Source License
+ * Version 2.0 (the 'License'). You may not use this file except in
+ * compliance with the License. The rights granted to you under the License
+ * may not be used to create, or enable the creation or redistribution of,
+ * unlawful or unlicensed copies of an Apple operating system, or to
+ * circumvent, violate, or enable the circumvention or violation of, any
+ * terms of an Apple operating system software license agreement.
+ *
+ * Please obtain a copy of the License at
+ * http://www.opensource.apple.com/apsl/ and read it before using this file.
+ *
+ * The Original Code and all software distributed under the License are
+ * distributed on an 'AS IS' basis, WITHOUT WARRANTY OF ANY KIND, EITHER
+ * EXPRESS OR IMPLIED, AND APPLE HEREBY DISCLAIMS ALL SUCH WARRANTIES,
+ * INCLUDING WITHOUT LIMITATION, ANY WARRANTIES OF MERCHANTABILITY,
+ * FITNESS FOR A PARTICULAR PURPOSE, QUIET ENJOYMENT OR NON-INFRINGEMENT.
+ * Please see the License for the specific language governing rights and
+ * limitations under the License.
+ *
+ * @APPLE_OSREFERENCE_LICENSE_HEADER_END@
+ */
+
+#ifdef __x86_64__
+#include <i386/mp.h>
+#include <i386/cpu_data.h>
+#include <i386/bit_routines.h>
+#include <i386/machine_cpu.h>
+#include <i386/machine_routines.h>
+#include <i386/misc_protos.h>
+#include <i386/serial_io.h>
+#endif /* __x86_64__ */
+
+#include <libkern/OSAtomic.h>
+#include <vm/vm_kern.h>
+#include <vm/vm_map.h>
+#include <console/video_console.h>
+#include <console/serial_protos.h>
+#include <kern/kalloc.h>
+#include <kern/thread.h>
+#include <kern/cpu_data.h>
+
+
+
+#ifndef MAX_CPU_SLOTS
+#define MAX_CPU_SLOTS (MAX_CPUS)
+#endif
+
+static struct {
+       char * buffer;
+       int len;
+       int used;
+       char * write_ptr;
+       char * read_ptr;
+       decl_simple_lock_data(, read_lock);
+       decl_simple_lock_data(, write_lock);
+} console_ring;
+
+hw_lock_data_t cnputc_lock;
+static volatile uint32_t console_output = 0;
+
+/*
+ * New allocation mechanism for console buffers
+ * Total allocation: 1 * PAGE_SIZE
+ * - Each cpu gets CPU_CONS_BUF_SIZE buffer
+ * - Kernel wide console ring gets PAGE_SIZE - MAX_CPU_SLOTS * CPU_CONS_BUF_SIZE
+ *
+ * At the return from console_init() the memory is setup as follows:
+ *  +----------------------------+-------------+-------------+-------------+-------------+
+ *  |console ring buffer---------|f2eec075-----|f2eec075-----|f2eec075-----|f2eec075-----|
+ *  +----------------------------+-------------+-------------+-------------+-------------+
+ * Each cpu allocation will find the first (f2eec075) and use that buffer.
+ *
+ */
+
+#define CPU_CONS_BUF_SIZE 256
+#define CPU_BUF_FREE_HEX 0xf2eec075
+
+#define KERN_CONSOLE_BUF_SIZE vm_map_round_page(CPU_CONS_BUF_SIZE *(MAX_CPU_SLOTS + 1), PAGE_SIZE - 1)
+#define KERN_CONSOLE_RING_SIZE (KERN_CONSOLE_BUF_SIZE - (CPU_CONS_BUF_SIZE * MAX_CPU_SLOTS))
+
+/*
+ * A serial line running at 115200 bps can output ~11.5 characters per millisecond.
+ * Synchronous serial logging with preemption+interrupts disabled fundamentally prevents us
+ * from hitting expected scheduling deadlines, but we can at least tone it down a bit.
+ *
+ * TODO: IOLog should use asynchronous serial logging instead of the synchronous serial console. (26555148)
+ *
+ * Keep interrupt disabled periods shorter than 1ms
+ */
+#define MAX_INT_DISABLED_FLUSH_SIZE 8
+#define MAX_TOTAL_FLUSH_SIZE (MAX(2, MAX_CPU_SLOTS) * CPU_CONS_BUF_SIZE)
+
+typedef struct console_buf {
+       char * buf_base;
+       char * buf_end;
+       char * buf_ptr;
+#define CPU_BUFFER_LEN (CPU_CONS_BUF_SIZE - 3 * (sizeof(char *)))
+       char buf[CPU_BUFFER_LEN];
+} console_buf_t;
+
+extern int serial_getc(void);
+extern void serial_putc(char);
+
+static void _serial_putc(int, int, int);
+
+struct console_ops cons_ops[] = {
+    {
+        .putc = _serial_putc, .getc = _serial_getc,
+    },
+    {
+        .putc = vcputc, .getc = vcgetc,
+    },
+};
+
+uint32_t nconsops = (sizeof cons_ops / sizeof cons_ops[0]);
+
+uint32_t cons_ops_index = VC_CONS_OPS;
+
+
+static bool console_suspended = false;
+
+static void
+console_ring_lock_init(void)
+{
+       simple_lock_init(&console_ring.read_lock, 0);
+       simple_lock_init(&console_ring.write_lock, 0);
+}
+
+void
+console_init(void)
+{
+       int ret, i;
+       uint32_t * p;
+
+       if (!OSCompareAndSwap(0, KERN_CONSOLE_RING_SIZE, (UInt32 *)&console_ring.len))
+               return;
+
+       assert(console_ring.len > 0);
+
+       ret = kmem_alloc(kernel_map, (vm_offset_t *)&console_ring.buffer, KERN_CONSOLE_BUF_SIZE, VM_KERN_MEMORY_OSFMK);
+       if (ret != KERN_SUCCESS) {
+               panic("console_ring_init() failed to allocate ring buffer, error %d\n", ret);
+       }
+
+       /* setup memory for per cpu console buffers */
+       for (i = 0; i < MAX_CPU_SLOTS; i++) {
+               p  = (uint32_t *)((uintptr_t)console_ring.buffer + console_ring.len + (i * sizeof(console_buf_t)));
+               *p = CPU_BUF_FREE_HEX;
+       }
+
+       console_ring.used      = 0;
+       console_ring.read_ptr  = console_ring.buffer;
+       console_ring.write_ptr = console_ring.buffer;
+       console_ring_lock_init();
+       hw_lock_init(&cnputc_lock);
+}
+
+void *
+console_cpu_alloc(__unused boolean_t boot_processor)
+{
+       console_buf_t * cbp;
+       int i;
+       uint32_t * p;
+
+       console_init();
+       assert(console_ring.buffer != NULL);
+
+       /* select the next slot from the per cpu buffers at end of console_ring.buffer */
+       for (i = 0; i < MAX_CPU_SLOTS; i++) {
+               p = (uint32_t *)((uintptr_t)console_ring.buffer + console_ring.len + (i * sizeof(console_buf_t)));
+               if (OSCompareAndSwap(CPU_BUF_FREE_HEX, 0, (UInt32 *)p))
+                       break;
+       }
+       assert(i < MAX_CPU_SLOTS);
+
+       cbp = (console_buf_t *)(uintptr_t)p;
+       if ((uintptr_t)cbp >= (uintptr_t)console_ring.buffer + KERN_CONSOLE_BUF_SIZE) {
+               printf("console_cpu_alloc() failed to allocate cpu buffer\n");
+               return NULL;
+       }
+
+       cbp->buf_base = (char *)&cbp->buf;
+       cbp->buf_ptr  = cbp->buf_base;
+       cbp->buf_end = cbp->buf_base + CPU_BUFFER_LEN;
+       return (void *)cbp;
+}
+
+void
+console_cpu_free(void * buf)
+{
+       assert((uintptr_t)buf > (uintptr_t)console_ring.buffer);
+       assert((uintptr_t)buf < (uintptr_t)console_ring.buffer + KERN_CONSOLE_BUF_SIZE);
+       if (buf != NULL)
+               *(uint32_t *)buf = CPU_BUF_FREE_HEX;
+}
+
+static inline int
+console_ring_space(void)
+{
+       return console_ring.len - console_ring.used;
+}
+
+static boolean_t
+console_ring_put(char ch)
+{
+       if (console_ring.used < console_ring.len) {
+               console_ring.used++;
+               *console_ring.write_ptr++ = ch;
+               if (console_ring.write_ptr - console_ring.buffer == console_ring.len)
+                       console_ring.write_ptr = console_ring.buffer;
+               return TRUE;
+       } else {
+               return FALSE;
+       }
+}
+
+static inline boolean_t
+cpu_buffer_put(console_buf_t * cbp, char ch)
+{
+       if (ch != '\0' && cbp->buf_ptr < cbp->buf_end) {
+               *(cbp->buf_ptr++) = ch;
+               return TRUE;
+       } else {
+               return FALSE;
+       }
+}
+
+static inline int
+cpu_buffer_size(console_buf_t * cbp)
+{
+       return (int)(cbp->buf_ptr - cbp->buf_base);
+}
+
+static inline void
+_cnputs(char * c, int size)
+{
+       /* The console device output routines are assumed to be
+        * non-reentrant.
+        */
+       mp_disable_preemption();
+       if (!hw_lock_to(&cnputc_lock, LockTimeOut)) {
+               /* If we timed out on the lock, and we're in the debugger,
+                * copy lock data for debugging and break the lock.
+                */
+               hw_lock_data_t _shadow_lock;
+               memcpy(&_shadow_lock, &cnputc_lock, sizeof(cnputc_lock));
+               if (debug_mode) {
+                       /* Since hw_lock_to takes a pre-emption count...*/
+                       mp_enable_preemption();
+                       hw_lock_init(&cnputc_lock);
+                       hw_lock_lock(&cnputc_lock);
+               } else {
+                       panic("Lock acquire timeout in _cnputs() lock=%p, lock owner thread=0x%lx, current_thread: %p\n", &_shadow_lock,
+                             _shadow_lock.lock_data, current_thread());
+               }
+       }
+
+       while (size-- > 0) {
+               cons_ops[cons_ops_index].putc(0, 0, *c);
+               if (*c == '\n')
+                       cons_ops[cons_ops_index].putc(0, 0, '\r');
+               c++;
+       }
+
+       hw_lock_unlock(&cnputc_lock);
+       mp_enable_preemption();
+}
+
+void
+cnputc_unbuffered(char c)
+{
+       _cnputs(&c, 1);
+}
+
+void
+cnputcusr(char c)
+{
+       boolean_t state;
+
+       /* Spin (with pre-emption enabled) waiting for console_ring_try_empty()
+        * to complete output. There is a small window here where we could
+        * end up with a stale value of console_output, but it's unlikely,
+        * and _cnputs(), which outputs to the console device, is internally
+        * synchronized. There's something of a conflict between the
+        * character-at-a-time (with pre-emption enabled) unbuffered
+        * output model here, and the buffered output from cnputc(),
+        * whose consumers include printf() ( which outputs a sequence
+        * with pre-emption disabled, and should be safe to call with
+        * interrupts off); we don't want to disable pre-emption indefinitely
+        * here, and spinlocks and mutexes are inappropriate.
+        */
+       while (console_output != 0)
+               ;
+
+       /*
+        * We disable interrupts to avoid issues caused by rendevous IPIs
+        * and an interruptible core holding the lock while an uninterruptible
+        * core wants it.  Stackshot is the prime example of this.
+        */
+       state = ml_set_interrupts_enabled(FALSE);
+       _cnputs(&c, 1);
+       ml_set_interrupts_enabled(state);
+}
+
+static void
+console_ring_try_empty(void)
+{
+#ifdef __x86_64__
+       boolean_t handle_tlb_flushes = (ml_get_interrupts_enabled() == FALSE);
+#endif /* __x86_64__ */
+
+       int nchars_out       = 0;
+       int total_chars_out  = 0;
+       int size_before_wrap = 0;
+
+       do {
+#ifdef __x86_64__
+               if (handle_tlb_flushes)
+                       handle_pending_TLB_flushes();
+#endif /* __x86_64__ */
+
+               /*
+                * Try to get the read lock on the ring buffer to empty it.
+                * If this fails someone else is already emptying...
+                */
+               if (!simple_lock_try(&console_ring.read_lock)) {
+                       /*
+                        * If multiple cores are spinning trying to empty the buffer,
+                        * we may suffer lock starvation (get the read lock, but
+                        * never the write lock, with other cores unable to get the
+                        * read lock).  As a result, insert a delay on failure, to
+                        * let other cores have a turn.
+                        */
+                       delay(1);
+                       return;
+               }
+
+               boolean_t state = ml_set_interrupts_enabled(FALSE);
+
+               /* Indicate that we're in the process of writing a block of data to the console. */
+               (void)hw_atomic_add(&console_output, 1);
+
+               simple_lock_try_lock_loop(&console_ring.write_lock);
+
+               /* try small chunk at a time, so we allow writes from other cpus into the buffer */
+               nchars_out = MIN(console_ring.used, MAX_INT_DISABLED_FLUSH_SIZE);
+
+               /* account for data to be read before wrap around */
+               size_before_wrap = (int)((console_ring.buffer + console_ring.len) - console_ring.read_ptr);
+               if (nchars_out > size_before_wrap)
+                       nchars_out = size_before_wrap;
+
+               if (nchars_out > 0) {
+                       _cnputs(console_ring.read_ptr, nchars_out);
+                       console_ring.read_ptr =
+                           console_ring.buffer + ((console_ring.read_ptr - console_ring.buffer + nchars_out) % console_ring.len);
+                       console_ring.used -= nchars_out;
+                       total_chars_out += nchars_out;
+               }
+
+               simple_unlock(&console_ring.write_lock);
+
+               (void)hw_atomic_sub(&console_output, 1);
+
+               simple_unlock(&console_ring.read_lock);
+
+               ml_set_interrupts_enabled(state);
+
+               /*
+                * In case we end up being the console drain thread
+                * for far too long, break out. Except in panic/suspend cases
+                * where we should clear out full buffer.
+                */
+               if (debug_mode == 0 && !console_suspended && (total_chars_out >= MAX_TOTAL_FLUSH_SIZE))
+                       break;
+
+       } while (nchars_out > 0);
+}
+
+
+void
+console_suspend()
+{
+       console_suspended = true;
+       console_ring_try_empty();
+}
+
+void
+console_resume()
+{
+       console_suspended = false;
+}
+
+void
+console_write(char * str, int size)
+{
+       console_init();
+       int chunk_size = size;
+       int i          = 0;
+
+       if (size > console_ring.len)
+               chunk_size = CPU_CONS_BUF_SIZE;
+
+       while (size > 0) {
+               boolean_t state = ml_set_interrupts_enabled(FALSE);
+
+               simple_lock_try_lock_loop(&console_ring.write_lock);
+               while (chunk_size > console_ring_space()) {
+                       simple_unlock(&console_ring.write_lock);
+                       ml_set_interrupts_enabled(state);
+
+                       console_ring_try_empty();
+
+                       state = ml_set_interrupts_enabled(FALSE);
+                       simple_lock_try_lock_loop(&console_ring.write_lock);
+               }
+
+               for (i = 0; i < chunk_size; i++)
+                       console_ring_put(str[i]);
+
+               str = &str[i];
+               size -= chunk_size;
+               simple_unlock(&console_ring.write_lock);
+               ml_set_interrupts_enabled(state);
+       }
+
+       console_ring_try_empty();
+}
+
+void
+cnputc(char c)
+{
+       console_buf_t * cbp;
+       cpu_data_t * cpu_data_p;
+       boolean_t state;
+       boolean_t needs_print = TRUE;
+       char * cp;
+
+restart:
+       mp_disable_preemption();
+       cpu_data_p = current_cpu_datap();
+       cbp = (console_buf_t *)cpu_data_p->cpu_console_buf;
+       if (console_suspended || cbp == NULL) {
+               mp_enable_preemption();
+               /* Put directly if console ring is not initialized or we're heading into suspend */
+               _cnputs(&c, 1);
+               return;
+       }
+
+#ifndef __x86_64__
+       /* Is there a panic backtrace going on? */
+       if (cpu_data_p->PAB_active) {
+               /* If another processor was in the process of emptying the
+                * console ring buffer when it received the panic backtrace
+                * signal, that processor will be spinning in DebugXCall()
+                * waiting for the panicking processor to finish printing
+                * the backtrace. But panicking processor will never
+                * be able to obtain the ring buffer lock since it is
+                * owned by a processor that's spinning in DebugXCall().
+                * Blow away any locks that other processors may have on
+                * the console ring buffer so that the backtrace can
+                * complete.
+                */
+               console_ring_lock_init();
+       }
+#endif /* __x86_64__ */
+
+       state = ml_set_interrupts_enabled(FALSE);
+
+       /*
+        * add to stack buf
+        * If the cpu buffer is full, we'll flush, then try
+        * another put.  If it fails a second time... screw
+        * it.
+        */
+       if (needs_print && !cpu_buffer_put(cbp, c)) {
+               simple_lock_try_lock_loop(&console_ring.write_lock);
+
+               if (cpu_buffer_size(cbp) > console_ring_space()) {
+                       simple_unlock(&console_ring.write_lock);
+                       ml_set_interrupts_enabled(state);
+                       mp_enable_preemption();
+
+                       console_ring_try_empty();
+                       goto restart;
+               }
+
+               for (cp = cbp->buf_base; cp < cbp->buf_ptr; cp++)
+                       console_ring_put(*cp);
+               cbp->buf_ptr = cbp->buf_base;
+               simple_unlock(&console_ring.write_lock);
+
+               cpu_buffer_put(cbp, c);
+       }
+
+       needs_print = FALSE;
+
+       if (c != '\n') {
+               ml_set_interrupts_enabled(state);
+               mp_enable_preemption();
+               return;
+       }
+
+       /* We printed a newline, time to flush the CPU buffer to the global buffer */
+       simple_lock_try_lock_loop(&console_ring.write_lock);
+
+       /*
+        * Is there enough space in the shared ring buffer?
+        * Try to empty if not.
+        * Note, we want the entire local buffer to fit to
+        * avoid another cpu interjecting.
+        */
+
+       if (cpu_buffer_size(cbp) > console_ring_space()) {
+               simple_unlock(&console_ring.write_lock);
+               ml_set_interrupts_enabled(state);
+               mp_enable_preemption();
+
+               console_ring_try_empty();
+
+               goto restart;
+       }
+
+       for (cp = cbp->buf_base; cp < cbp->buf_ptr; cp++)
+               console_ring_put(*cp);
+
+       cbp->buf_ptr = cbp->buf_base;
+       simple_unlock(&console_ring.write_lock);
+       ml_set_interrupts_enabled(state);
+       mp_enable_preemption();
+
+       console_ring_try_empty();
+
+       return;
+}
+
+int
+_serial_getc(__unused int a, __unused int b, boolean_t wait, __unused boolean_t raw)
+{
+       int c;
+       do {
+               c = serial_getc();
+       } while (wait && c < 0);
+
+
+       return c;
+}
+
+static void
+_serial_putc(__unused int a, __unused int b, int c)
+{
+       serial_putc(c);
+}
+
+int
+cngetc(void)
+{
+       return cons_ops[cons_ops_index].getc(0, 0, TRUE, FALSE);
+}
+
+int
+cnmaygetc(void)
+{
+       return cons_ops[cons_ops_index].getc(0, 0, FALSE, FALSE);
+}
+
+int
+vcgetc(__unused int l, __unused int u, __unused boolean_t wait, __unused boolean_t raw)
+{
+       char c;
+
+       if (0 == (*PE_poll_input)(0, &c))
+               return c;
+       else
+               return 0;
+}
+
+/* So we can re-write the serial device functions at boot-time */
+void
+console_set_serial_ops(struct console_ops * newops)
+{
+       cons_ops[SERIAL_CONS_OPS] = *newops;
+}
+