2 * Copyright (c) 1999, 2000, 2002-2008 Apple Inc. All rights reserved.
4 * @APPLE_LICENSE_HEADER_START@
6 * This file contains Original Code and/or Modifications of Original Code
7 * as defined in and that are subject to the Apple Public Source License
8 * Version 2.0 (the 'License'). You may not use this file except in
9 * compliance with the License. Please obtain a copy of the License at
10 * http://www.opensource.apple.com/apsl/ and read it before using this
13 * The Original Code and all software distributed under the License are
14 * distributed on an 'AS IS' basis, WITHOUT WARRANTY OF ANY KIND, EITHER
15 * EXPRESS OR IMPLIED, AND APPLE HEREBY DISCLAIMS ALL SUCH WARRANTIES,
16 * INCLUDING WITHOUT LIMITATION, ANY WARRANTIES OF MERCHANTABILITY,
17 * FITNESS FOR A PARTICULAR PURPOSE, QUIET ENJOYMENT OR NON-INFRINGEMENT.
18 * Please see the License for the specific language governing rights and
19 * limitations under the License.
21 * @APPLE_LICENSE_HEADER_END@
24 /* Bertrand from vmutils -> CF -> System */
26 #import "stack_logging.h"
27 #import "malloc_printf.h"
32 #include <mach/vm_statistics.h>
33 #import <malloc/malloc.h>
36 extern void spin_lock(int *);
37 extern void spin_unlock(int *);
38 extern void thread_stack_pcs(vm_address_t
*, unsigned, unsigned *);
39 extern const char *__crashreporter_info__
;
41 static inline void *allocate_pages(unsigned) __attribute__((always_inline
));
42 static inline void *allocate_pages(unsigned bytes
) {
44 if (vm_allocate(mach_task_self(), (vm_address_t
*)&address
, bytes
,
45 VM_MAKE_TAG(VM_MEMORY_ANALYSIS_TOOL
)| TRUE
)) {
46 malloc_printf("*** out of memory while stack logging\n");
47 __crashreporter_info__
= "*** out of memory while stack logging\n";
50 return (void *)address
;
53 static inline void deallocate_pages(void *, unsigned) __attribute__((always_inline
));
54 static inline void deallocate_pages(void *ptr
, unsigned bytes
) {
55 vm_deallocate(mach_task_self(), (vm_address_t
)ptr
, bytes
);
58 static inline void copy_pages(const void *, void *, unsigned) __attribute__((always_inline
));
59 static inline void copy_pages(const void *source
, void *dest
, unsigned bytes
) {
60 if (vm_copy(mach_task_self(), (vm_address_t
)source
, bytes
, (vm_address_t
)dest
)) memmove(dest
, source
, bytes
);
63 /*************** Uniquing stack ***********/
67 #define MAX_NUM_PC 512
69 static int enter_pair_in_table(unsigned *table
, unsigned numPages
, unsigned *uniquedParent
, unsigned thisPC
) {
70 // uniquedParent is in-out; return 1 is collisions max not exceeded
71 unsigned base
= numPages
* vm_page_size
/ (sizeof(int)*2*2);
72 unsigned hash
= base
+ (((*uniquedParent
) << 4) ^ (thisPC
>> 2)) % (base
- 1); // modulo odd number for hashing
73 unsigned collisions
= MAX_COLLIDE
;
74 while (collisions
--) {
75 unsigned *head
= table
+ hash
*2;
76 if (! head
[0] && !head
[1]) {
77 /* end of chain; store this entry! */
78 /* Note that we need to test for both head[0] and head[1] as (0, -1) is a valid entry */
80 head
[1] = *uniquedParent
;
81 *uniquedParent
= hash
;
84 if ((head
[0] == thisPC
) && (head
[1] == *uniquedParent
)) {
85 /* we found the proper entry, the value for the pair is the entry offset */
86 *uniquedParent
= hash
;
90 if (hash
== base
*2) hash
= base
;
95 unsigned stack_logging_get_unique_stack(unsigned **table
, unsigned *table_num_pages
, unsigned *stack_entries
, unsigned count
, unsigned num_hot_to_skip
) {
96 unsigned uniquedParent
= (unsigned)-1;
97 // we skip the warmest entries that are an artefact of the code
98 while (num_hot_to_skip
--) {
99 if (count
> 0) { stack_entries
++; count
--; }
102 unsigned thisPC
= stack_entries
[count
];
103 while (!enter_pair_in_table(*table
, *table_num_pages
, &uniquedParent
, thisPC
)) {
105 unsigned oldBytes
= (*table_num_pages
) * vm_page_size
;
106 newTable
= allocate_pages(oldBytes
*2);
107 copy_pages(*table
, newTable
, oldBytes
);
108 deallocate_pages(*table
, oldBytes
);
109 *table_num_pages
*= 2;
113 return uniquedParent
;
116 /*************** Logging stack and arguments ***********/
118 stack_logging_record_list_t
*stack_logging_the_record_list
= NULL
;
120 int stack_logging_enable_logging
= 0;
122 int stack_logging_dontcompact
= 0;
124 static int stack_logging_spin_lock
= 0;
126 static stack_logging_record_list_t
*GrowLogRecords(stack_logging_record_list_t
*records
, unsigned desiredNumRecords
) {
127 stack_logging_record_list_t
*new_records
;
128 unsigned old_size
= records
->overall_num_bytes
;
129 if (desiredNumRecords
*sizeof(stack_logging_record_t
)+sizeof(stack_logging_record_list_t
) < records
->overall_num_bytes
) return records
;
130 records
->overall_num_bytes
+= records
->overall_num_bytes
+ vm_page_size
; // in order to always get an even number of pages
131 new_records
= allocate_pages(records
->overall_num_bytes
);
132 copy_pages(records
, new_records
, old_size
);
133 deallocate_pages(records
, old_size
);
137 static void prepare_to_log_stack(void) {
138 if (!stack_logging_the_record_list
) {
139 unsigned totalSize
= 4 * vm_page_size
;
140 stack_logging_the_record_list
= allocate_pages(totalSize
);
141 memset(stack_logging_the_record_list
, 0, sizeof(stack_logging_record_list_t
));
142 stack_logging_the_record_list
->overall_num_bytes
= totalSize
;
143 stack_logging_the_record_list
->uniquing_table_num_pages
= 128;
144 stack_logging_the_record_list
->uniquing_table
= allocate_pages(stack_logging_the_record_list
->uniquing_table_num_pages
* vm_page_size
);
148 void stack_logging_log_stack(unsigned type
, unsigned arg1
, unsigned arg2
, unsigned arg3
, unsigned result
, unsigned num_hot_to_skip
) {
149 stack_logging_record_t
*rec
;
150 if (!stack_logging_enable_logging
) return;
151 // printf("stack_logging_log_stack 0x%x 0x%x 0x%x 0x%x -> 0x%x\n", type, arg1, arg2, arg3, result);
152 if (type
& stack_logging_flag_zone
) {
153 // just process it now and be done with it!
154 arg1
= arg2
; arg2
= arg3
; arg3
= 0; type
&= ~stack_logging_flag_zone
;
156 if (type
& stack_logging_flag_calloc
) {
157 // just process it now and be done with it!
158 arg1
*= arg2
; arg2
= arg3
; arg3
= 0; type
&= ~stack_logging_flag_calloc
;
160 if (type
& stack_logging_flag_object
) {
161 unsigned *class = (unsigned *)(long)arg1
;
162 arg1
= arg2
+ class[5]; // corresponds to the instance_size field
163 arg2
= 0; arg3
= 0; type
= stack_logging_type_alloc
;
165 if (type
& stack_logging_flag_cleared
) {
166 type
&= ~stack_logging_flag_cleared
;
168 if (type
& stack_logging_flag_handle
) {
169 if (stack_logging_type_alloc
) {
171 stack_logging_log_stack(stack_logging_type_alloc
, 0, 0, 0, result
, num_hot_to_skip
+1);
172 stack_logging_log_stack(stack_logging_type_alloc
, arg1
, 0, 0, *((int *)(long)result
), num_hot_to_skip
+1);
175 if (stack_logging_type_dealloc
) {
177 stack_logging_log_stack(stack_logging_type_dealloc
, *((int *)(long)arg1
), 0, 0, 0, num_hot_to_skip
+1);
178 stack_logging_log_stack(stack_logging_type_dealloc
, arg1
, 0, 0, 0, num_hot_to_skip
+1);
181 fprintf(stderr
, "*** Unknown logging type: 0x%x\n", type
);
183 if (type
== stack_logging_flag_set_handle_size
) {
185 // Thanks to a horrible hack, arg3 contains the prvious handle value
186 if (arg3
== *((int *)(long)arg1
)) return;
187 stack_logging_log_stack(stack_logging_type_dealloc
, arg3
, 0, 0, 0, num_hot_to_skip
+1);
188 stack_logging_log_stack(stack_logging_type_alloc
, arg2
, 0, 0, *((int *)(long)arg1
), num_hot_to_skip
+1);
191 if (type
== (stack_logging_type_dealloc
|stack_logging_type_alloc
)) {
192 if (arg1
== result
) return; // realloc had no effect, skipping
194 // realloc(NULL, size) same as malloc(size)
195 type
= stack_logging_type_alloc
; arg1
= arg2
; arg2
= arg3
; arg3
= 0;
197 // realloc(arg1, arg2) -> result is same as free(arg1); malloc(arg2) -> result
198 stack_logging_log_stack(stack_logging_type_dealloc
, arg1
, 0, 0, 0, num_hot_to_skip
+1);
199 stack_logging_log_stack(stack_logging_type_alloc
, arg2
, 0, 0, result
, num_hot_to_skip
+1);
203 if (type
== stack_logging_type_dealloc
) {
205 if (!arg1
) return; // free(nil)
207 prepare_to_log_stack();
208 spin_lock(&stack_logging_spin_lock
);
209 stack_logging_the_record_list
= GrowLogRecords(stack_logging_the_record_list
, stack_logging_the_record_list
->num_records
+ 1);
210 rec
= stack_logging_the_record_list
->records
+ stack_logging_the_record_list
->num_records
;
211 // We take care of the common case of alloc-dealloc
212 if (!stack_logging_dontcompact
&& stack_logging_the_record_list
->num_records
&& (type
== stack_logging_type_dealloc
) && arg1
&& ((rec
-1)->type
== stack_logging_type_alloc
) && (arg1
== STACK_LOGGING_DISGUISE((rec
-1)->address
))) {
213 stack_logging_the_record_list
->num_records
--;
214 // printf("Erased previous record in alloc-dealloc sequence\n");
216 unsigned stack_entries
[MAX_NUM_PC
];
219 if (type
== stack_logging_type_dealloc
) {
221 rec
->address
= STACK_LOGGING_DISGUISE(arg1
); // we disguise the address
222 } else if (type
== stack_logging_type_alloc
) {
223 rec
->argument
= arg1
;
224 rec
->address
= STACK_LOGGING_DISGUISE(result
); // we disguise the address
226 rec
->argument
= arg2
;
227 rec
->address
= STACK_LOGGING_DISGUISE(arg1
); // we disguise the address
229 // printf("Before getting samples 0x%x 0x%x 0x%x 0x%x -> 0x%x\n", type, arg1, arg2, arg3, result);
230 thread_stack_pcs((vm_address_t
*)stack_entries
, MAX_NUM_PC
- 1, &count
);
231 // We put at the bottom of the stack a marker that denotes the thread (+1 for good measure...)
232 stack_entries
[count
++] = (int)(long)pthread_self() + 1;
233 /* now let's unique the sample */
234 // printf("Uniquing 0x%x 0x%x 0x%x 0x%x -> 0x%x\n", type, arg1, arg2, arg3, result);
235 rec
->uniqued_stack
= stack_logging_get_unique_stack(&stack_logging_the_record_list
->uniquing_table
, &stack_logging_the_record_list
->uniquing_table_num_pages
, stack_entries
, count
, num_hot_to_skip
+2); // we additionally skip the warmest 2 entries that are an artefact of the code
236 stack_logging_the_record_list
->num_records
++;
238 spin_unlock(&stack_logging_spin_lock
);
241 static kern_return_t
default_reader(task_t task
, vm_address_t address
, vm_size_t size
, void **ptr
) {
242 *ptr
= (void *)address
;
246 static kern_return_t
get_remote_records(task_t task
, memory_reader_t reader
, stack_logging_record_list_t
**records
) {
248 vm_address_t
*remote_records_address_ref
;
251 err
= reader(task
, (vm_address_t
)&stack_logging_the_record_list
, sizeof(vm_address_t
), (void **)&remote_records_address_ref
);
253 if (!*remote_records_address_ref
) {
254 // printf("stack_logging: no stack record\n");
257 // printf("stack_logging: stack records at %p\n", (void *)(*remote_records_address_ref));
258 // printf("stack_logging: reading %d bytes\n", sizeof(stack_logging_record_list_t));
259 err
= reader(task
, *remote_records_address_ref
, sizeof(stack_logging_record_list_t
), (void **)records
); // get the list head
261 // printf("stack_logging: overall num bytes = %d\n", records->overall_num_bytes);
262 return reader(task
, *remote_records_address_ref
, (*records
)->overall_num_bytes
, (void **)records
);
265 kern_return_t
stack_logging_get_frames(task_t task
, memory_reader_t reader
, vm_address_t address
, vm_address_t
*stack_frames_buffer
, unsigned max_stack_frames
, unsigned *num_frames
) {
266 stack_logging_record_list_t
*records
;
269 unsigned disguised
= STACK_LOGGING_DISGUISE(address
);
270 if (!reader
) reader
= default_reader
;
272 err
= get_remote_records(task
, reader
, &records
);
273 if (err
|| !records
) return err
;
274 // printf("stack_logging: %d records\n", records->num_records);
276 while (index
< records
->num_records
) {
277 stack_logging_record_t
*record
= records
->records
+ index
;
278 if (record
->address
== disguised
) {
279 return stack_logging_frames_for_uniqued_stack(task
, reader
, record
->uniqued_stack
, stack_frames_buffer
, max_stack_frames
, num_frames
);
283 fprintf(stderr
, "*** stack_logging: no record found for 0x%lx\n", (long)address
);
287 kern_return_t
stack_logging_enumerate_records(task_t task
, memory_reader_t reader
, vm_address_t address
, void enumerator(stack_logging_record_t
, void *), void *context
) {
288 stack_logging_record_list_t
*records
;
291 unsigned disguised
= STACK_LOGGING_DISGUISE(address
);
292 if (!reader
) reader
= default_reader
;
293 err
= get_remote_records(task
, reader
, &records
);
294 if (err
|| !records
) return err
;
295 // printf("stack_logging: %d records\n", records->num_records);
297 while (index
< records
->num_records
) {
298 stack_logging_record_t
*record
= records
->records
+ index
;
299 if (!address
|| (record
->address
== disguised
)) enumerator(*record
, context
);
305 kern_return_t
stack_logging_frames_for_uniqued_stack(task_t task
, memory_reader_t reader
, unsigned uniqued_stack
, vm_address_t
*stack_frames_buffer
, unsigned max_stack_frames
, unsigned *num_frames
) {
306 stack_logging_record_list_t
*records
;
307 unsigned *uniquing_table
;
309 if (!reader
) reader
= default_reader
;
311 err
= get_remote_records(task
, reader
, &records
);
312 if (err
|| !records
) return err
;
313 err
= reader(task
, (vm_address_t
)records
->uniquing_table
, records
->uniquing_table_num_pages
* vm_page_size
, (void **)&uniquing_table
);
315 while (max_stack_frames
&& (uniqued_stack
!= -1)) {
317 if ((uniqued_stack
* 2 + 1) * sizeof(unsigned) >= records
->uniquing_table_num_pages
* vm_page_size
) {
318 fprintf(stderr
, "*** stack_logging: Invalid uniqued stack 0x%x", uniqued_stack
);
321 thisPC
= uniquing_table
[uniqued_stack
* 2];
322 uniqued_stack
= uniquing_table
[uniqued_stack
* 2 + 1];
323 if (!thisPC
&& !uniqued_stack
) {
325 fprintf(stderr
, "*** stack_logging: Invalid entry 0x%x", thisPC
);
328 stack_frames_buffer
[0] = thisPC
;
329 stack_frames_buffer
++;