]> git.saurik.com Git - redis.git/blame - deps/jemalloc.orig/src/prof.c
Jemalloc updated to 3.0.0.
[redis.git] / deps / jemalloc.orig / src / prof.c
CommitLineData
ad4c0b41 1#define JEMALLOC_PROF_C_
2#include "jemalloc/internal/jemalloc_internal.h"
3#ifdef JEMALLOC_PROF
4/******************************************************************************/
5
6#ifdef JEMALLOC_PROF_LIBUNWIND
7#define UNW_LOCAL_ONLY
8#include <libunwind.h>
9#endif
10
11#ifdef JEMALLOC_PROF_LIBGCC
12#include <unwind.h>
13#endif
14
15/******************************************************************************/
16/* Data. */
17
18bool opt_prof = false;
19bool opt_prof_active = true;
20size_t opt_lg_prof_bt_max = LG_PROF_BT_MAX_DEFAULT;
21size_t opt_lg_prof_sample = LG_PROF_SAMPLE_DEFAULT;
22ssize_t opt_lg_prof_interval = LG_PROF_INTERVAL_DEFAULT;
23bool opt_prof_gdump = false;
24bool opt_prof_leak = false;
25bool opt_prof_accum = true;
26ssize_t opt_lg_prof_tcmax = LG_PROF_TCMAX_DEFAULT;
27char opt_prof_prefix[PATH_MAX + 1];
28
29uint64_t prof_interval;
30bool prof_promote;
31
32unsigned prof_bt_max;
33
34#ifndef NO_TLS
35__thread prof_tdata_t *prof_tdata_tls
36 JEMALLOC_ATTR(tls_model("initial-exec"));
37#endif
38pthread_key_t prof_tdata_tsd;
39
40/*
41 * Global hash of (prof_bt_t *)-->(prof_ctx_t *). This is the master data
42 * structure that knows about all backtraces currently captured.
43 */
44static ckh_t bt2ctx;
45static malloc_mutex_t bt2ctx_mtx;
46
47static malloc_mutex_t prof_dump_seq_mtx;
48static uint64_t prof_dump_seq;
49static uint64_t prof_dump_iseq;
50static uint64_t prof_dump_mseq;
51static uint64_t prof_dump_useq;
52
53/*
54 * This buffer is rather large for stack allocation, so use a single buffer for
55 * all profile dumps. The buffer is implicitly protected by bt2ctx_mtx, since
56 * it must be locked anyway during dumping.
57 */
58static char prof_dump_buf[PROF_DUMP_BUF_SIZE];
59static unsigned prof_dump_buf_end;
60static int prof_dump_fd;
61
62/* Do not dump any profiles until bootstrapping is complete. */
63static bool prof_booted = false;
64
65static malloc_mutex_t enq_mtx;
66static bool enq;
67static bool enq_idump;
68static bool enq_gdump;
69
70/******************************************************************************/
71/* Function prototypes for non-inline static functions. */
72
73static prof_bt_t *bt_dup(prof_bt_t *bt);
74static void bt_destroy(prof_bt_t *bt);
75#ifdef JEMALLOC_PROF_LIBGCC
76static _Unwind_Reason_Code prof_unwind_init_callback(
77 struct _Unwind_Context *context, void *arg);
78static _Unwind_Reason_Code prof_unwind_callback(
79 struct _Unwind_Context *context, void *arg);
80#endif
81static bool prof_flush(bool propagate_err);
82static bool prof_write(const char *s, bool propagate_err);
83static void prof_ctx_sum(prof_ctx_t *ctx, prof_cnt_t *cnt_all,
84 size_t *leak_nctx);
85static void prof_ctx_destroy(prof_ctx_t *ctx);
86static void prof_ctx_merge(prof_ctx_t *ctx, prof_thr_cnt_t *cnt);
87static bool prof_dump_ctx(prof_ctx_t *ctx, prof_bt_t *bt,
88 bool propagate_err);
89static bool prof_dump_maps(bool propagate_err);
90static bool prof_dump(const char *filename, bool leakcheck,
91 bool propagate_err);
92static void prof_dump_filename(char *filename, char v, int64_t vseq);
93static void prof_fdump(void);
94static void prof_bt_hash(const void *key, unsigned minbits, size_t *hash1,
95 size_t *hash2);
96static bool prof_bt_keycomp(const void *k1, const void *k2);
97static void prof_tdata_cleanup(void *arg);
98
99/******************************************************************************/
100
101void
102bt_init(prof_bt_t *bt, void **vec)
103{
104
105 bt->vec = vec;
106 bt->len = 0;
107}
108
109static void
110bt_destroy(prof_bt_t *bt)
111{
112
113 idalloc(bt);
114}
115
116static prof_bt_t *
117bt_dup(prof_bt_t *bt)
118{
119 prof_bt_t *ret;
120
121 /*
122 * Create a single allocation that has space for vec immediately
123 * following the prof_bt_t structure. The backtraces that get
124 * stored in the backtrace caches are copied from stack-allocated
125 * temporary variables, so size is known at creation time. Making this
126 * a contiguous object improves cache locality.
127 */
128 ret = (prof_bt_t *)imalloc(QUANTUM_CEILING(sizeof(prof_bt_t)) +
129 (bt->len * sizeof(void *)));
130 if (ret == NULL)
131 return (NULL);
132 ret->vec = (void **)((uintptr_t)ret +
133 QUANTUM_CEILING(sizeof(prof_bt_t)));
134 memcpy(ret->vec, bt->vec, bt->len * sizeof(void *));
135 ret->len = bt->len;
136
137 return (ret);
138}
139
140static inline void
141prof_enter(void)
142{
143
144 malloc_mutex_lock(&enq_mtx);
145 enq = true;
146 malloc_mutex_unlock(&enq_mtx);
147
148 malloc_mutex_lock(&bt2ctx_mtx);
149}
150
151static inline void
152prof_leave(void)
153{
154 bool idump, gdump;
155
156 malloc_mutex_unlock(&bt2ctx_mtx);
157
158 malloc_mutex_lock(&enq_mtx);
159 enq = false;
160 idump = enq_idump;
161 enq_idump = false;
162 gdump = enq_gdump;
163 enq_gdump = false;
164 malloc_mutex_unlock(&enq_mtx);
165
166 if (idump)
167 prof_idump();
168 if (gdump)
169 prof_gdump();
170}
171
172#ifdef JEMALLOC_PROF_LIBUNWIND
173void
174prof_backtrace(prof_bt_t *bt, unsigned nignore, unsigned max)
175{
176 unw_context_t uc;
177 unw_cursor_t cursor;
178 unsigned i;
179 int err;
180
181 assert(bt->len == 0);
182 assert(bt->vec != NULL);
183 assert(max <= (1U << opt_lg_prof_bt_max));
184
185 unw_getcontext(&uc);
186 unw_init_local(&cursor, &uc);
187
188 /* Throw away (nignore+1) stack frames, if that many exist. */
189 for (i = 0; i < nignore + 1; i++) {
190 err = unw_step(&cursor);
191 if (err <= 0)
192 return;
193 }
194
195 /*
196 * Iterate over stack frames until there are no more, or until no space
197 * remains in bt.
198 */
199 for (i = 0; i < max; i++) {
200 unw_get_reg(&cursor, UNW_REG_IP, (unw_word_t *)&bt->vec[i]);
201 bt->len++;
202 err = unw_step(&cursor);
203 if (err <= 0)
204 break;
205 }
206}
207#endif
208#ifdef JEMALLOC_PROF_LIBGCC
209static _Unwind_Reason_Code
210prof_unwind_init_callback(struct _Unwind_Context *context, void *arg)
211{
212
213 return (_URC_NO_REASON);
214}
215
216static _Unwind_Reason_Code
217prof_unwind_callback(struct _Unwind_Context *context, void *arg)
218{
219 prof_unwind_data_t *data = (prof_unwind_data_t *)arg;
220
221 if (data->nignore > 0)
222 data->nignore--;
223 else {
224 data->bt->vec[data->bt->len] = (void *)_Unwind_GetIP(context);
225 data->bt->len++;
226 if (data->bt->len == data->max)
227 return (_URC_END_OF_STACK);
228 }
229
230 return (_URC_NO_REASON);
231}
232
233void
234prof_backtrace(prof_bt_t *bt, unsigned nignore, unsigned max)
235{
236 prof_unwind_data_t data = {bt, nignore, max};
237
238 _Unwind_Backtrace(prof_unwind_callback, &data);
239}
240#endif
241#ifdef JEMALLOC_PROF_GCC
242void
243prof_backtrace(prof_bt_t *bt, unsigned nignore, unsigned max)
244{
245#define BT_FRAME(i) \
246 if ((i) < nignore + max) { \
247 void *p; \
248 if (__builtin_frame_address(i) == 0) \
249 return; \
250 p = __builtin_return_address(i); \
251 if (p == NULL) \
252 return; \
253 if (i >= nignore) { \
254 bt->vec[(i) - nignore] = p; \
255 bt->len = (i) - nignore + 1; \
256 } \
257 } else \
258 return;
259
260 assert(nignore <= 3);
261 assert(max <= (1U << opt_lg_prof_bt_max));
262
263 BT_FRAME(0)
264 BT_FRAME(1)
265 BT_FRAME(2)
266 BT_FRAME(3)
267 BT_FRAME(4)
268 BT_FRAME(5)
269 BT_FRAME(6)
270 BT_FRAME(7)
271 BT_FRAME(8)
272 BT_FRAME(9)
273
274 BT_FRAME(10)
275 BT_FRAME(11)
276 BT_FRAME(12)
277 BT_FRAME(13)
278 BT_FRAME(14)
279 BT_FRAME(15)
280 BT_FRAME(16)
281 BT_FRAME(17)
282 BT_FRAME(18)
283 BT_FRAME(19)
284
285 BT_FRAME(20)
286 BT_FRAME(21)
287 BT_FRAME(22)
288 BT_FRAME(23)
289 BT_FRAME(24)
290 BT_FRAME(25)
291 BT_FRAME(26)
292 BT_FRAME(27)
293 BT_FRAME(28)
294 BT_FRAME(29)
295
296 BT_FRAME(30)
297 BT_FRAME(31)
298 BT_FRAME(32)
299 BT_FRAME(33)
300 BT_FRAME(34)
301 BT_FRAME(35)
302 BT_FRAME(36)
303 BT_FRAME(37)
304 BT_FRAME(38)
305 BT_FRAME(39)
306
307 BT_FRAME(40)
308 BT_FRAME(41)
309 BT_FRAME(42)
310 BT_FRAME(43)
311 BT_FRAME(44)
312 BT_FRAME(45)
313 BT_FRAME(46)
314 BT_FRAME(47)
315 BT_FRAME(48)
316 BT_FRAME(49)
317
318 BT_FRAME(50)
319 BT_FRAME(51)
320 BT_FRAME(52)
321 BT_FRAME(53)
322 BT_FRAME(54)
323 BT_FRAME(55)
324 BT_FRAME(56)
325 BT_FRAME(57)
326 BT_FRAME(58)
327 BT_FRAME(59)
328
329 BT_FRAME(60)
330 BT_FRAME(61)
331 BT_FRAME(62)
332 BT_FRAME(63)
333 BT_FRAME(64)
334 BT_FRAME(65)
335 BT_FRAME(66)
336 BT_FRAME(67)
337 BT_FRAME(68)
338 BT_FRAME(69)
339
340 BT_FRAME(70)
341 BT_FRAME(71)
342 BT_FRAME(72)
343 BT_FRAME(73)
344 BT_FRAME(74)
345 BT_FRAME(75)
346 BT_FRAME(76)
347 BT_FRAME(77)
348 BT_FRAME(78)
349 BT_FRAME(79)
350
351 BT_FRAME(80)
352 BT_FRAME(81)
353 BT_FRAME(82)
354 BT_FRAME(83)
355 BT_FRAME(84)
356 BT_FRAME(85)
357 BT_FRAME(86)
358 BT_FRAME(87)
359 BT_FRAME(88)
360 BT_FRAME(89)
361
362 BT_FRAME(90)
363 BT_FRAME(91)
364 BT_FRAME(92)
365 BT_FRAME(93)
366 BT_FRAME(94)
367 BT_FRAME(95)
368 BT_FRAME(96)
369 BT_FRAME(97)
370 BT_FRAME(98)
371 BT_FRAME(99)
372
373 BT_FRAME(100)
374 BT_FRAME(101)
375 BT_FRAME(102)
376 BT_FRAME(103)
377 BT_FRAME(104)
378 BT_FRAME(105)
379 BT_FRAME(106)
380 BT_FRAME(107)
381 BT_FRAME(108)
382 BT_FRAME(109)
383
384 BT_FRAME(110)
385 BT_FRAME(111)
386 BT_FRAME(112)
387 BT_FRAME(113)
388 BT_FRAME(114)
389 BT_FRAME(115)
390 BT_FRAME(116)
391 BT_FRAME(117)
392 BT_FRAME(118)
393 BT_FRAME(119)
394
395 BT_FRAME(120)
396 BT_FRAME(121)
397 BT_FRAME(122)
398 BT_FRAME(123)
399 BT_FRAME(124)
400 BT_FRAME(125)
401 BT_FRAME(126)
402 BT_FRAME(127)
403
404 /* Extras to compensate for nignore. */
405 BT_FRAME(128)
406 BT_FRAME(129)
407 BT_FRAME(130)
408#undef BT_FRAME
409}
410#endif
411
412prof_thr_cnt_t *
413prof_lookup(prof_bt_t *bt)
414{
415 union {
416 prof_thr_cnt_t *p;
417 void *v;
418 } ret;
419 prof_tdata_t *prof_tdata;
420
421 prof_tdata = PROF_TCACHE_GET();
422 if (prof_tdata == NULL) {
423 prof_tdata = prof_tdata_init();
424 if (prof_tdata == NULL)
425 return (NULL);
426 }
427
428 if (ckh_search(&prof_tdata->bt2cnt, bt, NULL, &ret.v)) {
429 union {
430 prof_bt_t *p;
431 void *v;
432 } btkey;
433 union {
434 prof_ctx_t *p;
435 void *v;
436 } ctx;
437 bool new_ctx;
438
439 /*
440 * This thread's cache lacks bt. Look for it in the global
441 * cache.
442 */
443 prof_enter();
444 if (ckh_search(&bt2ctx, bt, &btkey.v, &ctx.v)) {
445 /* bt has never been seen before. Insert it. */
446 ctx.v = imalloc(sizeof(prof_ctx_t));
447 if (ctx.v == NULL) {
448 prof_leave();
449 return (NULL);
450 }
451 btkey.p = bt_dup(bt);
452 if (btkey.v == NULL) {
453 prof_leave();
454 idalloc(ctx.v);
455 return (NULL);
456 }
457 ctx.p->bt = btkey.p;
458 if (malloc_mutex_init(&ctx.p->lock)) {
459 prof_leave();
460 idalloc(btkey.v);
461 idalloc(ctx.v);
462 return (NULL);
463 }
464 memset(&ctx.p->cnt_merged, 0, sizeof(prof_cnt_t));
465 ql_new(&ctx.p->cnts_ql);
466 if (ckh_insert(&bt2ctx, btkey.v, ctx.v)) {
467 /* OOM. */
468 prof_leave();
469 malloc_mutex_destroy(&ctx.p->lock);
470 idalloc(btkey.v);
471 idalloc(ctx.v);
472 return (NULL);
473 }
474 /*
475 * Artificially raise curobjs, in order to avoid a race
476 * condition with prof_ctx_merge()/prof_ctx_destroy().
477 *
478 * No locking is necessary for ctx here because no other
479 * threads have had the opportunity to fetch it from
480 * bt2ctx yet.
481 */
482 ctx.p->cnt_merged.curobjs++;
483 new_ctx = true;
484 } else {
485 /*
486 * Artificially raise curobjs, in order to avoid a race
487 * condition with prof_ctx_merge()/prof_ctx_destroy().
488 */
489 malloc_mutex_lock(&ctx.p->lock);
490 ctx.p->cnt_merged.curobjs++;
491 malloc_mutex_unlock(&ctx.p->lock);
492 new_ctx = false;
493 }
494 prof_leave();
495
496 /* Link a prof_thd_cnt_t into ctx for this thread. */
497 if (opt_lg_prof_tcmax >= 0 && ckh_count(&prof_tdata->bt2cnt)
498 == (ZU(1) << opt_lg_prof_tcmax)) {
499 assert(ckh_count(&prof_tdata->bt2cnt) > 0);
500 /*
501 * Flush the least recently used cnt in order to keep
502 * bt2cnt from becoming too large.
503 */
504 ret.p = ql_last(&prof_tdata->lru_ql, lru_link);
505 assert(ret.v != NULL);
506 if (ckh_remove(&prof_tdata->bt2cnt, ret.p->ctx->bt,
507 NULL, NULL))
508 assert(false);
509 ql_remove(&prof_tdata->lru_ql, ret.p, lru_link);
510 prof_ctx_merge(ret.p->ctx, ret.p);
511 /* ret can now be re-used. */
512 } else {
513 assert(opt_lg_prof_tcmax < 0 ||
514 ckh_count(&prof_tdata->bt2cnt) < (ZU(1) <<
515 opt_lg_prof_tcmax));
516 /* Allocate and partially initialize a new cnt. */
517 ret.v = imalloc(sizeof(prof_thr_cnt_t));
518 if (ret.p == NULL) {
519 if (new_ctx)
520 prof_ctx_destroy(ctx.p);
521 return (NULL);
522 }
523 ql_elm_new(ret.p, cnts_link);
524 ql_elm_new(ret.p, lru_link);
525 }
526 /* Finish initializing ret. */
527 ret.p->ctx = ctx.p;
528 ret.p->epoch = 0;
529 memset(&ret.p->cnts, 0, sizeof(prof_cnt_t));
530 if (ckh_insert(&prof_tdata->bt2cnt, btkey.v, ret.v)) {
531 if (new_ctx)
532 prof_ctx_destroy(ctx.p);
533 idalloc(ret.v);
534 return (NULL);
535 }
536 ql_head_insert(&prof_tdata->lru_ql, ret.p, lru_link);
537 malloc_mutex_lock(&ctx.p->lock);
538 ql_tail_insert(&ctx.p->cnts_ql, ret.p, cnts_link);
539 ctx.p->cnt_merged.curobjs--;
540 malloc_mutex_unlock(&ctx.p->lock);
541 } else {
542 /* Move ret to the front of the LRU. */
543 ql_remove(&prof_tdata->lru_ql, ret.p, lru_link);
544 ql_head_insert(&prof_tdata->lru_ql, ret.p, lru_link);
545 }
546
547 return (ret.p);
548}
549
550static bool
551prof_flush(bool propagate_err)
552{
553 bool ret = false;
554 ssize_t err;
555
556 err = write(prof_dump_fd, prof_dump_buf, prof_dump_buf_end);
557 if (err == -1) {
558 if (propagate_err == false) {
559 malloc_write("<jemalloc>: write() failed during heap "
560 "profile flush\n");
561 if (opt_abort)
562 abort();
563 }
564 ret = true;
565 }
566 prof_dump_buf_end = 0;
567
568 return (ret);
569}
570
571static bool
572prof_write(const char *s, bool propagate_err)
573{
574 unsigned i, slen, n;
575
576 i = 0;
577 slen = strlen(s);
578 while (i < slen) {
579 /* Flush the buffer if it is full. */
580 if (prof_dump_buf_end == PROF_DUMP_BUF_SIZE)
581 if (prof_flush(propagate_err) && propagate_err)
582 return (true);
583
584 if (prof_dump_buf_end + slen <= PROF_DUMP_BUF_SIZE) {
585 /* Finish writing. */
586 n = slen - i;
587 } else {
588 /* Write as much of s as will fit. */
589 n = PROF_DUMP_BUF_SIZE - prof_dump_buf_end;
590 }
591 memcpy(&prof_dump_buf[prof_dump_buf_end], &s[i], n);
592 prof_dump_buf_end += n;
593 i += n;
594 }
595
596 return (false);
597}
598
599static void
600prof_ctx_sum(prof_ctx_t *ctx, prof_cnt_t *cnt_all, size_t *leak_nctx)
601{
602 prof_thr_cnt_t *thr_cnt;
603 prof_cnt_t tcnt;
604
605 malloc_mutex_lock(&ctx->lock);
606
607 memcpy(&ctx->cnt_summed, &ctx->cnt_merged, sizeof(prof_cnt_t));
608 ql_foreach(thr_cnt, &ctx->cnts_ql, cnts_link) {
609 volatile unsigned *epoch = &thr_cnt->epoch;
610
611 while (true) {
612 unsigned epoch0 = *epoch;
613
614 /* Make sure epoch is even. */
615 if (epoch0 & 1U)
616 continue;
617
618 memcpy(&tcnt, &thr_cnt->cnts, sizeof(prof_cnt_t));
619
620 /* Terminate if epoch didn't change while reading. */
621 if (*epoch == epoch0)
622 break;
623 }
624
625 ctx->cnt_summed.curobjs += tcnt.curobjs;
626 ctx->cnt_summed.curbytes += tcnt.curbytes;
627 if (opt_prof_accum) {
628 ctx->cnt_summed.accumobjs += tcnt.accumobjs;
629 ctx->cnt_summed.accumbytes += tcnt.accumbytes;
630 }
631 }
632
633 if (ctx->cnt_summed.curobjs != 0)
634 (*leak_nctx)++;
635
636 /* Add to cnt_all. */
637 cnt_all->curobjs += ctx->cnt_summed.curobjs;
638 cnt_all->curbytes += ctx->cnt_summed.curbytes;
639 if (opt_prof_accum) {
640 cnt_all->accumobjs += ctx->cnt_summed.accumobjs;
641 cnt_all->accumbytes += ctx->cnt_summed.accumbytes;
642 }
643
644 malloc_mutex_unlock(&ctx->lock);
645}
646
647static void
648prof_ctx_destroy(prof_ctx_t *ctx)
649{
650
651 /*
652 * Check that ctx is still unused by any thread cache before destroying
653 * it. prof_lookup() artificially raises ctx->cnt_merge.curobjs in
654 * order to avoid a race condition with this function, as does
655 * prof_ctx_merge() in order to avoid a race between the main body of
656 * prof_ctx_merge() and entry into this function.
657 */
658 prof_enter();
659 malloc_mutex_lock(&ctx->lock);
660 if (ql_first(&ctx->cnts_ql) == NULL && ctx->cnt_merged.curobjs == 1) {
661 assert(ctx->cnt_merged.curbytes == 0);
662 assert(ctx->cnt_merged.accumobjs == 0);
663 assert(ctx->cnt_merged.accumbytes == 0);
664 /* Remove ctx from bt2ctx. */
665 if (ckh_remove(&bt2ctx, ctx->bt, NULL, NULL))
666 assert(false);
667 prof_leave();
668 /* Destroy ctx. */
669 malloc_mutex_unlock(&ctx->lock);
670 bt_destroy(ctx->bt);
671 malloc_mutex_destroy(&ctx->lock);
672 idalloc(ctx);
673 } else {
674 /*
675 * Compensate for increment in prof_ctx_merge() or
676 * prof_lookup().
677 */
678 ctx->cnt_merged.curobjs--;
679 malloc_mutex_unlock(&ctx->lock);
680 prof_leave();
681 }
682}
683
684static void
685prof_ctx_merge(prof_ctx_t *ctx, prof_thr_cnt_t *cnt)
686{
687 bool destroy;
688
689 /* Merge cnt stats and detach from ctx. */
690 malloc_mutex_lock(&ctx->lock);
691 ctx->cnt_merged.curobjs += cnt->cnts.curobjs;
692 ctx->cnt_merged.curbytes += cnt->cnts.curbytes;
693 ctx->cnt_merged.accumobjs += cnt->cnts.accumobjs;
694 ctx->cnt_merged.accumbytes += cnt->cnts.accumbytes;
695 ql_remove(&ctx->cnts_ql, cnt, cnts_link);
696 if (opt_prof_accum == false && ql_first(&ctx->cnts_ql) == NULL &&
697 ctx->cnt_merged.curobjs == 0) {
698 /*
699 * Artificially raise ctx->cnt_merged.curobjs in order to keep
700 * another thread from winning the race to destroy ctx while
701 * this one has ctx->lock dropped. Without this, it would be
702 * possible for another thread to:
703 *
704 * 1) Sample an allocation associated with ctx.
705 * 2) Deallocate the sampled object.
706 * 3) Successfully prof_ctx_destroy(ctx).
707 *
708 * The result would be that ctx no longer exists by the time
709 * this thread accesses it in prof_ctx_destroy().
710 */
711 ctx->cnt_merged.curobjs++;
712 destroy = true;
713 } else
714 destroy = false;
715 malloc_mutex_unlock(&ctx->lock);
716 if (destroy)
717 prof_ctx_destroy(ctx);
718}
719
720static bool
721prof_dump_ctx(prof_ctx_t *ctx, prof_bt_t *bt, bool propagate_err)
722{
723 char buf[UMAX2S_BUFSIZE];
724 unsigned i;
725
726 if (opt_prof_accum == false && ctx->cnt_summed.curobjs == 0) {
727 assert(ctx->cnt_summed.curbytes == 0);
728 assert(ctx->cnt_summed.accumobjs == 0);
729 assert(ctx->cnt_summed.accumbytes == 0);
730 return (false);
731 }
732
733 if (prof_write(u2s(ctx->cnt_summed.curobjs, 10, buf), propagate_err)
734 || prof_write(": ", propagate_err)
735 || prof_write(u2s(ctx->cnt_summed.curbytes, 10, buf),
736 propagate_err)
737 || prof_write(" [", propagate_err)
738 || prof_write(u2s(ctx->cnt_summed.accumobjs, 10, buf),
739 propagate_err)
740 || prof_write(": ", propagate_err)
741 || prof_write(u2s(ctx->cnt_summed.accumbytes, 10, buf),
742 propagate_err)
743 || prof_write("] @", propagate_err))
744 return (true);
745
746 for (i = 0; i < bt->len; i++) {
747 if (prof_write(" 0x", propagate_err)
748 || prof_write(u2s((uintptr_t)bt->vec[i], 16, buf),
749 propagate_err))
750 return (true);
751 }
752
753 if (prof_write("\n", propagate_err))
754 return (true);
755
756 return (false);
757}
758
759static bool
760prof_dump_maps(bool propagate_err)
761{
762 int mfd;
763 char buf[UMAX2S_BUFSIZE];
764 char *s;
765 unsigned i, slen;
766 /* /proc/<pid>/maps\0 */
767 char mpath[6 + UMAX2S_BUFSIZE
768 + 5 + 1];
769
770 i = 0;
771
772 s = "/proc/";
773 slen = strlen(s);
774 memcpy(&mpath[i], s, slen);
775 i += slen;
776
777 s = u2s(getpid(), 10, buf);
778 slen = strlen(s);
779 memcpy(&mpath[i], s, slen);
780 i += slen;
781
782 s = "/maps";
783 slen = strlen(s);
784 memcpy(&mpath[i], s, slen);
785 i += slen;
786
787 mpath[i] = '\0';
788
789 mfd = open(mpath, O_RDONLY);
790 if (mfd != -1) {
791 ssize_t nread;
792
793 if (prof_write("\nMAPPED_LIBRARIES:\n", propagate_err) &&
794 propagate_err)
795 return (true);
796 nread = 0;
797 do {
798 prof_dump_buf_end += nread;
799 if (prof_dump_buf_end == PROF_DUMP_BUF_SIZE) {
800 /* Make space in prof_dump_buf before read(). */
801 if (prof_flush(propagate_err) && propagate_err)
802 return (true);
803 }
804 nread = read(mfd, &prof_dump_buf[prof_dump_buf_end],
805 PROF_DUMP_BUF_SIZE - prof_dump_buf_end);
806 } while (nread > 0);
807 close(mfd);
808 } else
809 return (true);
810
811 return (false);
812}
813
814static bool
815prof_dump(const char *filename, bool leakcheck, bool propagate_err)
816{
817 prof_cnt_t cnt_all;
818 size_t tabind;
819 union {
820 prof_bt_t *p;
821 void *v;
822 } bt;
823 union {
824 prof_ctx_t *p;
825 void *v;
826 } ctx;
827 char buf[UMAX2S_BUFSIZE];
828 size_t leak_nctx;
829
830 prof_enter();
831 prof_dump_fd = creat(filename, 0644);
832 if (prof_dump_fd == -1) {
833 if (propagate_err == false) {
834 malloc_write("<jemalloc>: creat(\"");
835 malloc_write(filename);
836 malloc_write("\", 0644) failed\n");
837 if (opt_abort)
838 abort();
839 }
840 goto ERROR;
841 }
842
843 /* Merge per thread profile stats, and sum them in cnt_all. */
844 memset(&cnt_all, 0, sizeof(prof_cnt_t));
845 leak_nctx = 0;
846 for (tabind = 0; ckh_iter(&bt2ctx, &tabind, NULL, &ctx.v) == false;)
847 prof_ctx_sum(ctx.p, &cnt_all, &leak_nctx);
848
849 /* Dump profile header. */
850 if (prof_write("heap profile: ", propagate_err)
851 || prof_write(u2s(cnt_all.curobjs, 10, buf), propagate_err)
852 || prof_write(": ", propagate_err)
853 || prof_write(u2s(cnt_all.curbytes, 10, buf), propagate_err)
854 || prof_write(" [", propagate_err)
855 || prof_write(u2s(cnt_all.accumobjs, 10, buf), propagate_err)
856 || prof_write(": ", propagate_err)
857 || prof_write(u2s(cnt_all.accumbytes, 10, buf), propagate_err))
858 goto ERROR;
859
860 if (opt_lg_prof_sample == 0) {
861 if (prof_write("] @ heapprofile\n", propagate_err))
862 goto ERROR;
863 } else {
864 if (prof_write("] @ heap_v2/", propagate_err)
865 || prof_write(u2s((uint64_t)1U << opt_lg_prof_sample, 10,
866 buf), propagate_err)
867 || prof_write("\n", propagate_err))
868 goto ERROR;
869 }
870
871 /* Dump per ctx profile stats. */
872 for (tabind = 0; ckh_iter(&bt2ctx, &tabind, &bt.v, &ctx.v)
873 == false;) {
874 if (prof_dump_ctx(ctx.p, bt.p, propagate_err))
875 goto ERROR;
876 }
877
878 /* Dump /proc/<pid>/maps if possible. */
879 if (prof_dump_maps(propagate_err))
880 goto ERROR;
881
882 if (prof_flush(propagate_err))
883 goto ERROR;
884 close(prof_dump_fd);
885 prof_leave();
886
887 if (leakcheck && cnt_all.curbytes != 0) {
888 malloc_write("<jemalloc>: Leak summary: ");
889 malloc_write(u2s(cnt_all.curbytes, 10, buf));
890 malloc_write((cnt_all.curbytes != 1) ? " bytes, " : " byte, ");
891 malloc_write(u2s(cnt_all.curobjs, 10, buf));
892 malloc_write((cnt_all.curobjs != 1) ? " objects, " :
893 " object, ");
894 malloc_write(u2s(leak_nctx, 10, buf));
895 malloc_write((leak_nctx != 1) ? " contexts\n" : " context\n");
896 malloc_write("<jemalloc>: Run pprof on \"");
897 malloc_write(filename);
898 malloc_write("\" for leak detail\n");
899 }
900
901 return (false);
902ERROR:
903 prof_leave();
904 return (true);
905}
906
907#define DUMP_FILENAME_BUFSIZE (PATH_MAX+ UMAX2S_BUFSIZE \
908 + 1 \
909 + UMAX2S_BUFSIZE \
910 + 2 \
911 + UMAX2S_BUFSIZE \
912 + 5 + 1)
913static void
914prof_dump_filename(char *filename, char v, int64_t vseq)
915{
916 char buf[UMAX2S_BUFSIZE];
917 char *s;
918 unsigned i, slen;
919
920 /*
921 * Construct a filename of the form:
922 *
923 * <prefix>.<pid>.<seq>.v<vseq>.heap\0
924 */
925
926 i = 0;
927
928 s = opt_prof_prefix;
929 slen = strlen(s);
930 memcpy(&filename[i], s, slen);
931 i += slen;
932
933 s = ".";
934 slen = strlen(s);
935 memcpy(&filename[i], s, slen);
936 i += slen;
937
938 s = u2s(getpid(), 10, buf);
939 slen = strlen(s);
940 memcpy(&filename[i], s, slen);
941 i += slen;
942
943 s = ".";
944 slen = strlen(s);
945 memcpy(&filename[i], s, slen);
946 i += slen;
947
948 s = u2s(prof_dump_seq, 10, buf);
949 prof_dump_seq++;
950 slen = strlen(s);
951 memcpy(&filename[i], s, slen);
952 i += slen;
953
954 s = ".";
955 slen = strlen(s);
956 memcpy(&filename[i], s, slen);
957 i += slen;
958
959 filename[i] = v;
960 i++;
961
962 if (vseq != 0xffffffffffffffffLLU) {
963 s = u2s(vseq, 10, buf);
964 slen = strlen(s);
965 memcpy(&filename[i], s, slen);
966 i += slen;
967 }
968
969 s = ".heap";
970 slen = strlen(s);
971 memcpy(&filename[i], s, slen);
972 i += slen;
973
974 filename[i] = '\0';
975}
976
977static void
978prof_fdump(void)
979{
980 char filename[DUMP_FILENAME_BUFSIZE];
981
982 if (prof_booted == false)
983 return;
984
985 if (opt_prof_prefix[0] != '\0') {
986 malloc_mutex_lock(&prof_dump_seq_mtx);
987 prof_dump_filename(filename, 'f', 0xffffffffffffffffLLU);
988 malloc_mutex_unlock(&prof_dump_seq_mtx);
989 prof_dump(filename, opt_prof_leak, false);
990 }
991}
992
993void
994prof_idump(void)
995{
996 char filename[DUMP_FILENAME_BUFSIZE];
997
998 if (prof_booted == false)
999 return;
1000 malloc_mutex_lock(&enq_mtx);
1001 if (enq) {
1002 enq_idump = true;
1003 malloc_mutex_unlock(&enq_mtx);
1004 return;
1005 }
1006 malloc_mutex_unlock(&enq_mtx);
1007
1008 if (opt_prof_prefix[0] != '\0') {
1009 malloc_mutex_lock(&prof_dump_seq_mtx);
1010 prof_dump_filename(filename, 'i', prof_dump_iseq);
1011 prof_dump_iseq++;
1012 malloc_mutex_unlock(&prof_dump_seq_mtx);
1013 prof_dump(filename, false, false);
1014 }
1015}
1016
1017bool
1018prof_mdump(const char *filename)
1019{
1020 char filename_buf[DUMP_FILENAME_BUFSIZE];
1021
1022 if (opt_prof == false || prof_booted == false)
1023 return (true);
1024
1025 if (filename == NULL) {
1026 /* No filename specified, so automatically generate one. */
1027 if (opt_prof_prefix[0] == '\0')
1028 return (true);
1029 malloc_mutex_lock(&prof_dump_seq_mtx);
1030 prof_dump_filename(filename_buf, 'm', prof_dump_mseq);
1031 prof_dump_mseq++;
1032 malloc_mutex_unlock(&prof_dump_seq_mtx);
1033 filename = filename_buf;
1034 }
1035 return (prof_dump(filename, false, true));
1036}
1037
1038void
1039prof_gdump(void)
1040{
1041 char filename[DUMP_FILENAME_BUFSIZE];
1042
1043 if (prof_booted == false)
1044 return;
1045 malloc_mutex_lock(&enq_mtx);
1046 if (enq) {
1047 enq_gdump = true;
1048 malloc_mutex_unlock(&enq_mtx);
1049 return;
1050 }
1051 malloc_mutex_unlock(&enq_mtx);
1052
1053 if (opt_prof_prefix[0] != '\0') {
1054 malloc_mutex_lock(&prof_dump_seq_mtx);
1055 prof_dump_filename(filename, 'u', prof_dump_useq);
1056 prof_dump_useq++;
1057 malloc_mutex_unlock(&prof_dump_seq_mtx);
1058 prof_dump(filename, false, false);
1059 }
1060}
1061
1062static void
1063prof_bt_hash(const void *key, unsigned minbits, size_t *hash1, size_t *hash2)
1064{
1065 size_t ret1, ret2;
1066 uint64_t h;
1067 prof_bt_t *bt = (prof_bt_t *)key;
1068
1069 assert(minbits <= 32 || (SIZEOF_PTR == 8 && minbits <= 64));
1070 assert(hash1 != NULL);
1071 assert(hash2 != NULL);
1072
1073 h = hash(bt->vec, bt->len * sizeof(void *), 0x94122f335b332aeaLLU);
1074 if (minbits <= 32) {
1075 /*
1076 * Avoid doing multiple hashes, since a single hash provides
1077 * enough bits.
1078 */
1079 ret1 = h & ZU(0xffffffffU);
1080 ret2 = h >> 32;
1081 } else {
1082 ret1 = h;
1083 ret2 = hash(bt->vec, bt->len * sizeof(void *),
1084 0x8432a476666bbc13LLU);
1085 }
1086
1087 *hash1 = ret1;
1088 *hash2 = ret2;
1089}
1090
1091static bool
1092prof_bt_keycomp(const void *k1, const void *k2)
1093{
1094 const prof_bt_t *bt1 = (prof_bt_t *)k1;
1095 const prof_bt_t *bt2 = (prof_bt_t *)k2;
1096
1097 if (bt1->len != bt2->len)
1098 return (false);
1099 return (memcmp(bt1->vec, bt2->vec, bt1->len * sizeof(void *)) == 0);
1100}
1101
1102prof_tdata_t *
1103prof_tdata_init(void)
1104{
1105 prof_tdata_t *prof_tdata;
1106
1107 /* Initialize an empty cache for this thread. */
1108 prof_tdata = (prof_tdata_t *)imalloc(sizeof(prof_tdata_t));
1109 if (prof_tdata == NULL)
1110 return (NULL);
1111
1112 if (ckh_new(&prof_tdata->bt2cnt, PROF_CKH_MINITEMS,
1113 prof_bt_hash, prof_bt_keycomp)) {
1114 idalloc(prof_tdata);
1115 return (NULL);
1116 }
1117 ql_new(&prof_tdata->lru_ql);
1118
1119 prof_tdata->vec = imalloc(sizeof(void *) * prof_bt_max);
1120 if (prof_tdata->vec == NULL) {
1121 ckh_delete(&prof_tdata->bt2cnt);
1122 idalloc(prof_tdata);
1123 return (NULL);
1124 }
1125
1126 prof_tdata->prn_state = 0;
1127 prof_tdata->threshold = 0;
1128 prof_tdata->accum = 0;
1129
1130 PROF_TCACHE_SET(prof_tdata);
1131
1132 return (prof_tdata);
1133}
1134
1135static void
1136prof_tdata_cleanup(void *arg)
1137{
1138 prof_thr_cnt_t *cnt;
1139 prof_tdata_t *prof_tdata = (prof_tdata_t *)arg;
1140
1141 /*
1142 * Delete the hash table. All of its contents can still be iterated
1143 * over via the LRU.
1144 */
1145 ckh_delete(&prof_tdata->bt2cnt);
1146
1147 /* Iteratively merge cnt's into the global stats and delete them. */
1148 while ((cnt = ql_last(&prof_tdata->lru_ql, lru_link)) != NULL) {
1149 ql_remove(&prof_tdata->lru_ql, cnt, lru_link);
1150 prof_ctx_merge(cnt->ctx, cnt);
1151 idalloc(cnt);
1152 }
1153
1154 idalloc(prof_tdata->vec);
1155
1156 idalloc(prof_tdata);
1157 PROF_TCACHE_SET(NULL);
1158}
1159
1160void
1161prof_boot0(void)
1162{
1163
1164 memcpy(opt_prof_prefix, PROF_PREFIX_DEFAULT,
1165 sizeof(PROF_PREFIX_DEFAULT));
1166}
1167
1168void
1169prof_boot1(void)
1170{
1171
1172 /*
1173 * opt_prof and prof_promote must be in their final state before any
1174 * arenas are initialized, so this function must be executed early.
1175 */
1176
1177 if (opt_prof_leak && opt_prof == false) {
1178 /*
1179 * Enable opt_prof, but in such a way that profiles are never
1180 * automatically dumped.
1181 */
1182 opt_prof = true;
1183 opt_prof_gdump = false;
1184 prof_interval = 0;
1185 } else if (opt_prof) {
1186 if (opt_lg_prof_interval >= 0) {
1187 prof_interval = (((uint64_t)1U) <<
1188 opt_lg_prof_interval);
1189 } else
1190 prof_interval = 0;
1191 }
1192
1193 prof_promote = (opt_prof && opt_lg_prof_sample > PAGE_SHIFT);
1194}
1195
1196bool
1197prof_boot2(void)
1198{
1199
1200 if (opt_prof) {
1201 if (ckh_new(&bt2ctx, PROF_CKH_MINITEMS, prof_bt_hash,
1202 prof_bt_keycomp))
1203 return (true);
1204 if (malloc_mutex_init(&bt2ctx_mtx))
1205 return (true);
1206 if (pthread_key_create(&prof_tdata_tsd, prof_tdata_cleanup)
1207 != 0) {
1208 malloc_write(
1209 "<jemalloc>: Error in pthread_key_create()\n");
1210 abort();
1211 }
1212
1213 prof_bt_max = (1U << opt_lg_prof_bt_max);
1214 if (malloc_mutex_init(&prof_dump_seq_mtx))
1215 return (true);
1216
1217 if (malloc_mutex_init(&enq_mtx))
1218 return (true);
1219 enq = false;
1220 enq_idump = false;
1221 enq_gdump = false;
1222
1223 if (atexit(prof_fdump) != 0) {
1224 malloc_write("<jemalloc>: Error in atexit()\n");
1225 if (opt_abort)
1226 abort();
1227 }
1228 }
1229
1230#ifdef JEMALLOC_PROF_LIBGCC
1231 /*
1232 * Cause the backtracing machinery to allocate its internal state
1233 * before enabling profiling.
1234 */
1235 _Unwind_Backtrace(prof_unwind_init_callback, NULL);
1236#endif
1237
1238 prof_booted = true;
1239
1240 return (false);
1241}
1242
1243/******************************************************************************/
1244#endif /* JEMALLOC_PROF */