2 * Copyright (c) 2005-2008 Apple Computer, Inc. All rights reserved.
4 * @APPLE_OSREFERENCE_LICENSE_HEADER_START@
6 * This file contains Original Code and/or Modifications of Original Code
7 * as defined in and that are subject to the Apple Public Source License
8 * Version 2.0 (the 'License'). You may not use this file except in
9 * compliance with the License. The rights granted to you under the License
10 * may not be used to create, or enable the creation or redistribution of,
11 * unlawful or unlicensed copies of an Apple operating system, or to
12 * circumvent, violate, or enable the circumvention or violation of, any
13 * terms of an Apple operating system software license agreement.
15 * Please obtain a copy of the License at
16 * http://www.opensource.apple.com/apsl/ and read it before using this file.
18 * The Original Code and all software distributed under the License are
19 * distributed on an 'AS IS' basis, WITHOUT WARRANTY OF ANY KIND, EITHER
20 * EXPRESS OR IMPLIED, AND APPLE HEREBY DISCLAIMS ALL SUCH WARRANTIES,
21 * INCLUDING WITHOUT LIMITATION, ANY WARRANTIES OF MERCHANTABILITY,
22 * FITNESS FOR A PARTICULAR PURPOSE, QUIET ENJOYMENT OR NON-INFRINGEMENT.
23 * Please see the License for the specific language governing rights and
24 * limitations under the License.
26 * @APPLE_OSREFERENCE_LICENSE_HEADER_END@
29 #define MACH__POSIX_C_SOURCE_PRIVATE 1 /* pulls in suitable savearea from
30 * mach/ppc/thread_status.h */
31 #include <arm/caches_internal.h>
32 #include <arm/proc_reg.h>
34 #include <kern/thread.h>
35 #include <mach/thread_status.h>
39 #include <sys/malloc.h>
41 #include <sys/systm.h>
43 #include <sys/proc_internal.h>
44 #include <sys/kauth.h>
45 #include <sys/dtrace.h>
46 #include <sys/dtrace_impl.h>
47 #include <libkern/OSAtomic.h>
48 #include <kern/simple_lock.h>
49 #include <kern/sched_prim.h> /* for thread_wakeup() */
50 #include <kern/thread_call.h>
51 #include <kern/task.h>
52 #include <miscfs/devfs/devfs.h>
53 #include <mach/vm_param.h>
55 extern struct arm_saved_state
*find_kern_regs(thread_t
);
57 extern dtrace_id_t dtrace_probeid_error
; /* special ERROR probe */
58 typedef arm_saved_state_t savearea_t
;
60 extern lck_attr_t
*dtrace_lck_attr
;
61 extern lck_grp_t
*dtrace_lck_grp
;
63 int dtrace_arm_condition_true(int condition
, int cpsr
);
66 * Atomicity and synchronization
69 dtrace_membar_producer(void)
72 __asm__
volatile ("dmb ish" : : : "memory");
74 __asm__
volatile ("nop" : : : "memory");
79 dtrace_membar_consumer(void)
82 __asm__
volatile ("dmb ish" : : : "memory");
84 __asm__
volatile ("nop" : : : "memory");
89 * Interrupt manipulation
90 * XXX dtrace_getipl() can be called from probe context.
96 * XXX Drat, get_interrupt_level is MACH_KERNEL_PRIVATE
97 * in osfmk/kern/cpu_data.h
99 /* return get_interrupt_level(); */
100 return ml_at_interrupt_context() ? 1 : 0;
108 decl_lck_mtx_data(static, dt_xc_lock
);
109 static uint32_t dt_xc_sync
;
111 typedef struct xcArg
{
120 xcArg_t
*pArg
= (xcArg_t
*) foo
;
122 if (pArg
->cpu
== CPU
->cpu_id
|| pArg
->cpu
== DTRACE_CPUALL
) {
123 (pArg
->f
)(pArg
->arg
);
126 if (hw_atomic_sub(&dt_xc_sync
, 1) == 0) {
127 thread_wakeup((event_t
) &dt_xc_sync
);
133 * dtrace_xcall() is not called from probe context.
136 dtrace_xcall(processorid_t cpu
, dtrace_xcall_t f
, void *arg
)
139 /* Only one dtrace_xcall in flight allowed */
140 lck_mtx_lock(&dt_xc_lock
);
148 cpu_broadcast_xcall(&dt_xc_sync
, TRUE
, xcRemote
, (void*) &xcArg
);
150 lck_mtx_unlock(&dt_xc_lock
);
154 /* On uniprocessor systems, the cpu should always be either ourselves or all */
155 ASSERT(cpu
== CPU
->cpu_id
|| cpu
== DTRACE_CPUALL
);
166 dtrace_isa_init(void)
169 lck_mtx_init(&dt_xc_lock
, dtrace_lck_grp
, dtrace_lck_attr
);
178 dtrace_getreg(struct regs
* savearea
, uint_t reg
)
180 struct arm_saved_state
*regs
= (struct arm_saved_state
*) savearea
;
182 DTRACE_CPUFLAG_SET(CPU_DTRACE_ILLOP
);
185 /* beyond register limit? */
186 if (reg
> ARM_SAVED_STATE32_COUNT
- 1) {
187 DTRACE_CPUFLAG_SET(CPU_DTRACE_ILLOP
);
191 return (uint64_t) ((unsigned int *) (&(regs
->r
)))[reg
];
194 #define RETURN_OFFSET 4
197 dtrace_getustack_common(uint64_t * pcstack
, int pcstack_limit
, user_addr_t pc
,
202 ASSERT(pcstack
== NULL
|| pcstack_limit
> 0);
206 if (pcstack
!= NULL
) {
207 *pcstack
++ = (uint64_t) pc
;
209 if (pcstack_limit
<= 0) {
218 pc
= dtrace_fuword32((sp
+ RETURN_OFFSET
));
219 sp
= dtrace_fuword32(sp
);
226 dtrace_getupcstack(uint64_t * pcstack
, int pcstack_limit
)
228 thread_t thread
= current_thread();
231 volatile uint16_t *flags
= (volatile uint16_t *) &cpu_core
[CPU
->cpu_id
].cpuc_dtrace_flags
;
234 if (*flags
& CPU_DTRACE_FAULT
) {
238 if (pcstack_limit
<= 0) {
243 * If there's no user context we still need to zero the stack.
245 if (thread
== NULL
) {
249 regs
= (savearea_t
*) find_user_regs(thread
);
254 *pcstack
++ = (uint64_t)dtrace_proc_selfpid();
257 if (pcstack_limit
<= 0) {
264 if (DTRACE_CPUFLAG_ISSET(CPU_DTRACE_ENTRY
)) {
265 *pcstack
++ = (uint64_t) pc
;
267 if (pcstack_limit
<= 0) {
274 n
= dtrace_getustack_common(pcstack
, pcstack_limit
, pc
, regs
->r
[7]);
277 ASSERT(n
<= pcstack_limit
);
283 while (pcstack_limit
-- > 0) {
289 dtrace_getustackdepth(void)
291 thread_t thread
= current_thread();
296 if (thread
== NULL
) {
300 if (DTRACE_CPUFLAG_ISSET(CPU_DTRACE_FAULT
)) {
304 regs
= (savearea_t
*) find_user_regs(thread
);
312 if (DTRACE_CPUFLAG_ISSET(CPU_DTRACE_ENTRY
)) {
318 * Note that unlike ppc, the arm code does not use
319 * CPU_DTRACE_USTACK_FP. This is because arm always
320 * traces from the sp, even in syscall/profile/fbt
324 n
+= dtrace_getustack_common(NULL
, 0, pc
, regs
->r
[7]);
330 dtrace_getufpstack(uint64_t * pcstack
, uint64_t * fpstack
, int pcstack_limit
)
332 /* XXX ARMTODO 64vs32 */
333 thread_t thread
= current_thread();
337 volatile uint16_t *flags
= (volatile uint16_t *) &cpu_core
[CPU
->cpu_id
].cpuc_dtrace_flags
;
340 uintptr_t oldcontext
;
344 if (*flags
& CPU_DTRACE_FAULT
) {
348 if (pcstack_limit
<= 0) {
353 * If there's no user context we still need to zero the stack.
355 if (thread
== NULL
) {
359 regs
= (savearea_t
*) find_user_regs(thread
);
364 *pcstack
++ = (uint64_t)dtrace_proc_selfpid();
367 if (pcstack_limit
<= 0) {
374 #if 0 /* XXX signal stack crawl */
375 oldcontext
= lwp
->lwp_oldcontext
;
377 if (p
->p_model
== DATAMODEL_NATIVE
) {
378 s1
= sizeof(struct frame
) + 2 * sizeof(long);
379 s2
= s1
+ sizeof(siginfo_t
);
381 s1
= sizeof(struct frame32
) + 3 * sizeof(int);
382 s2
= s1
+ sizeof(siginfo32_t
);
386 if (DTRACE_CPUFLAG_ISSET(CPU_DTRACE_ENTRY
)) {
387 *pcstack
++ = (uint64_t) pc
;
390 if (pcstack_limit
<= 0) {
394 pc
= dtrace_fuword32(sp
);
396 while (pc
!= 0 && sp
!= 0) {
397 *pcstack
++ = (uint64_t) pc
;
400 if (pcstack_limit
<= 0) {
404 #if 0 /* XXX signal stack crawl */
405 if (oldcontext
== sp
+ s1
|| oldcontext
== sp
+ s2
) {
406 if (p
->p_model
== DATAMODEL_NATIVE
) {
407 ucontext_t
*ucp
= (ucontext_t
*) oldcontext
;
408 greg_t
*gregs
= ucp
->uc_mcontext
.gregs
;
410 sp
= dtrace_fulword(&gregs
[REG_FP
]);
411 pc
= dtrace_fulword(&gregs
[REG_PC
]);
413 oldcontext
= dtrace_fulword(&ucp
->uc_link
);
415 ucontext_t
*ucp
= (ucontext_t
*) oldcontext
;
416 greg_t
*gregs
= ucp
->uc_mcontext
.gregs
;
418 sp
= dtrace_fuword32(&gregs
[EBP
]);
419 pc
= dtrace_fuword32(&gregs
[EIP
]);
421 oldcontext
= dtrace_fuword32(&ucp
->uc_link
);
426 pc
= dtrace_fuword32((sp
+ RETURN_OFFSET
));
427 sp
= dtrace_fuword32(sp
);
433 * This is totally bogus: if we faulted, we're going to clear
434 * the fault and break. This is to deal with the apparently
435 * broken Java stacks on x86.
437 if (*flags
& CPU_DTRACE_FAULT
) {
438 *flags
&= ~CPU_DTRACE_FAULT
;
445 while (pcstack_limit
-- > 0) {
451 dtrace_getpcstack(pc_t
* pcstack
, int pcstack_limit
, int aframes
,
454 struct frame
*fp
= (struct frame
*) __builtin_frame_address(0);
455 struct frame
*nextfp
, *minfp
, *stacktop
;
460 uintptr_t caller
= CPU
->cpu_dtrace_caller
;
462 if ((on_intr
= CPU_ON_INTR(CPU
)) != 0) {
463 stacktop
= (struct frame
*) dtrace_get_cpu_int_stack_top();
465 stacktop
= (struct frame
*) (dtrace_get_kernel_stack(current_thread()) + kernel_stack_size
);
472 if (intrpc
!= NULL
&& depth
< pcstack_limit
) {
473 pcstack
[depth
++] = (pc_t
) intrpc
;
476 while (depth
< pcstack_limit
) {
477 nextfp
= *(struct frame
**) fp
;
478 pc
= *(uintptr_t *) (((uint32_t) fp
) + RETURN_OFFSET
);
480 if (nextfp
<= minfp
|| nextfp
>= stacktop
) {
483 * Hop from interrupt stack to thread stack.
485 arm_saved_state_t
*arm_kern_regs
= (arm_saved_state_t
*) find_kern_regs(current_thread());
487 nextfp
= (struct frame
*)arm_kern_regs
->r
[7];
489 vm_offset_t kstack_base
= dtrace_get_kernel_stack(current_thread());
491 minfp
= (struct frame
*)kstack_base
;
492 stacktop
= (struct frame
*)(kstack_base
+ kernel_stack_size
);
496 if (nextfp
<= minfp
|| nextfp
>= stacktop
) {
501 * If this thread was on the interrupt stack, but did not
502 * take an interrupt (i.e, the idle thread), there is no
503 * explicit saved state for us to use.
509 * This is the last frame we can process; indicate
510 * that we should return after processing this frame.
516 if (--aframes
== 0 && caller
!= (uintptr_t)NULL
) {
518 * We've just run out of artificial frames,
519 * and we have a valid caller -- fill it in
522 ASSERT(depth
< pcstack_limit
);
523 pcstack
[depth
++] = (pc_t
) caller
;
524 caller
= (uintptr_t)NULL
;
527 if (depth
< pcstack_limit
) {
528 pcstack
[depth
++] = (pc_t
) pc
;
533 while (depth
< pcstack_limit
) {
534 pcstack
[depth
++] = (pc_t
) NULL
;
544 dtrace_instr_size(uint32_t instr
, int thumb_mode
)
547 uint16_t instr16
= *(uint16_t*) &instr
;
548 if (((instr16
>> 11) & 0x1F) > 0x1C) {
559 dtrace_getarg(int arg
, int aframes
, dtrace_mstate_t
*mstate
, dtrace_vstate_t
*vstate
)
561 #pragma unused(arg, aframes, mstate, vstate)
565 uintptr_t *fp
= (uintptr_t *)__builtin_frame_address(0);
570 for (i
= 1; i
<= aframes
; i
++) {
574 if (dtrace_invop_callsite_pre
!= NULL
575 && pc
> (uintptr_t)dtrace_invop_callsite_pre
576 && pc
<= (uintptr_t)dtrace_invop_callsite_post
) {
578 * If we pass through the invalid op handler, we will
579 * use the pointer that it passed to the stack as the
580 * second argument to dtrace_invop() as the pointer to
581 * the frame we're hunting for.
584 stack
= (uintptr_t *)&fp
[1]; /* Find marshalled arguments */
585 fp
= (struct frame
*)stack
[1]; /* Grab *second* argument */
586 stack
= (uintptr_t *)&fp
[1]; /* Find marshalled arguments */
587 DTRACE_CPUFLAG_SET(CPU_DTRACE_NOFAULT
);
588 val
= (uint64_t)(stack
[arg
]);
589 DTRACE_CPUFLAG_CLEAR(CPU_DTRACE_NOFAULT
);
595 * Arrive here when provider has called dtrace_probe directly.
597 stack
= (uintptr_t *)&fp
[1]; /* Find marshalled arguments */
598 stack
++; /* Advance past probeID */
600 DTRACE_CPUFLAG_SET(CPU_DTRACE_NOFAULT
);
601 val
= *(((uint64_t *)stack
) + arg
); /* dtrace_probe arguments arg0 .. arg4 are 64bits wide */
602 DTRACE_CPUFLAG_CLEAR(CPU_DTRACE_NOFAULT
);
605 return 0xfeedfacedeafbeadLL
;
609 dtrace_probe_error(dtrace_state_t
*state
, dtrace_epid_t epid
, int which
,
610 int fltoffs
, int fault
, uint64_t illval
)
614 * For the case of the error probe firing lets
615 * stash away "illval" here, and special-case retrieving it in DIF_VARIABLE_ARG.
617 state
->dts_arg_error_illval
= illval
;
618 dtrace_probe( dtrace_probeid_error
, (uint64_t)(uintptr_t)state
, epid
, which
, fltoffs
, fault
);
622 dtrace_toxic_ranges(void (*func
)(uintptr_t base
, uintptr_t limit
))
624 /* XXX ARMTODO check copied from ppc/x86*/
626 * "base" is the smallest toxic address in the range, "limit" is the first
627 * VALID address greater than "base".
629 func(0x0, VM_MIN_KERNEL_ADDRESS
);
630 if (VM_MAX_KERNEL_ADDRESS
< ~(uintptr_t)0) {
631 func(VM_MAX_KERNEL_ADDRESS
+ 1, ~(uintptr_t)0);
636 dtrace_arm_condition_true(int cond
, int cpsr
)
639 int zf
= (cpsr
& PSR_ZF
) ? 1 : 0,
640 nf
= (cpsr
& PSR_NF
) ? 1 : 0,
641 cf
= (cpsr
& PSR_CF
) ? 1 : 0,
642 vf
= (cpsr
& PSR_VF
) ? 1 : 0;
645 case 0: taken
= zf
; break;
646 case 1: taken
= !zf
; break;
647 case 2: taken
= cf
; break;
648 case 3: taken
= !cf
; break;
649 case 4: taken
= nf
; break;
650 case 5: taken
= !nf
; break;
651 case 6: taken
= vf
; break;
652 case 7: taken
= !vf
; break;
653 case 8: taken
= (cf
&& !zf
); break;
654 case 9: taken
= (!cf
|| zf
); break;
655 case 10: taken
= (nf
== vf
); break;
656 case 11: taken
= (nf
!= vf
); break;
657 case 12: taken
= (!zf
&& (nf
== vf
)); break;
658 case 13: taken
= (zf
|| (nf
!= vf
)); break;
659 case 14: taken
= 1; break;
660 case 15: taken
= 1; break; /* always "true" for ARM, unpredictable for THUMB. */
667 dtrace_flush_caches(void)
669 /* TODO There were some problems with flushing just the cache line that had been modified.
670 * For now, we'll flush the entire cache, until we figure out how to flush just the patched block.
673 InvalidatePoU_Icache();