1#ifndef JEMALLOC_INTERNAL_PROF_INLINES_B_H
2#define JEMALLOC_INTERNAL_PROF_INLINES_B_H
3
4#include "jemalloc/internal/safety_check.h"
5#include "jemalloc/internal/sz.h"
6
7JEMALLOC_ALWAYS_INLINE bool
8prof_gdump_get_unlocked(void) {
9	/*
10	 * No locking is used when reading prof_gdump_val in the fast path, so
11	 * there are no guarantees regarding how long it will take for all
12	 * threads to notice state changes.
13	 */
14	return prof_gdump_val;
15}
16
17JEMALLOC_ALWAYS_INLINE prof_tdata_t *
18prof_tdata_get(tsd_t *tsd, bool create) {
19	prof_tdata_t *tdata;
20
21	cassert(config_prof);
22
23	tdata = tsd_prof_tdata_get(tsd);
24	if (create) {
25		if (unlikely(tdata == NULL)) {
26			if (tsd_nominal(tsd)) {
27				tdata = prof_tdata_init(tsd);
28				tsd_prof_tdata_set(tsd, tdata);
29			}
30		} else if (unlikely(tdata->expired)) {
31			tdata = prof_tdata_reinit(tsd, tdata);
32			tsd_prof_tdata_set(tsd, tdata);
33		}
34		assert(tdata == NULL || tdata->attached);
35	}
36
37	return tdata;
38}
39
40JEMALLOC_ALWAYS_INLINE prof_tctx_t *
41prof_tctx_get(tsdn_t *tsdn, const void *ptr, alloc_ctx_t *alloc_ctx) {
42	cassert(config_prof);
43	assert(ptr != NULL);
44
45	return arena_prof_tctx_get(tsdn, ptr, alloc_ctx);
46}
47
48JEMALLOC_ALWAYS_INLINE void
49prof_tctx_set(tsdn_t *tsdn, const void *ptr, size_t usize,
50    alloc_ctx_t *alloc_ctx, prof_tctx_t *tctx) {
51	cassert(config_prof);
52	assert(ptr != NULL);
53
54	arena_prof_tctx_set(tsdn, ptr, usize, alloc_ctx, tctx);
55}
56
57JEMALLOC_ALWAYS_INLINE void
58prof_tctx_reset(tsdn_t *tsdn, const void *ptr, prof_tctx_t *tctx) {
59	cassert(config_prof);
60	assert(ptr != NULL);
61
62	arena_prof_tctx_reset(tsdn, ptr, tctx);
63}
64
65JEMALLOC_ALWAYS_INLINE nstime_t
66prof_alloc_time_get(tsdn_t *tsdn, const void *ptr, alloc_ctx_t *alloc_ctx) {
67	cassert(config_prof);
68	assert(ptr != NULL);
69
70	return arena_prof_alloc_time_get(tsdn, ptr, alloc_ctx);
71}
72
73JEMALLOC_ALWAYS_INLINE void
74prof_alloc_time_set(tsdn_t *tsdn, const void *ptr, alloc_ctx_t *alloc_ctx,
75    nstime_t t) {
76	cassert(config_prof);
77	assert(ptr != NULL);
78
79	arena_prof_alloc_time_set(tsdn, ptr, alloc_ctx, t);
80}
81
82JEMALLOC_ALWAYS_INLINE bool
83prof_sample_check(tsd_t *tsd, size_t usize, bool update) {
84	ssize_t check = update ? 0 : usize;
85
86	int64_t bytes_until_sample = tsd_bytes_until_sample_get(tsd);
87	if (update) {
88		bytes_until_sample -= usize;
89		if (tsd_nominal(tsd)) {
90			tsd_bytes_until_sample_set(tsd, bytes_until_sample);
91		}
92	}
93	if (likely(bytes_until_sample >= check)) {
94		return true;
95	}
96
97	return false;
98}
99
100JEMALLOC_ALWAYS_INLINE bool
101prof_sample_accum_update(tsd_t *tsd, size_t usize, bool update,
102			 prof_tdata_t **tdata_out) {
103	prof_tdata_t *tdata;
104
105	cassert(config_prof);
106
107	/* Fastpath: no need to load tdata */
108	if (likely(prof_sample_check(tsd, usize, update))) {
109		return true;
110	}
111
112	bool booted = tsd_prof_tdata_get(tsd);
113	tdata = prof_tdata_get(tsd, true);
114	if (unlikely((uintptr_t)tdata <= (uintptr_t)PROF_TDATA_STATE_MAX)) {
115		tdata = NULL;
116	}
117
118	if (tdata_out != NULL) {
119		*tdata_out = tdata;
120	}
121
122	if (unlikely(tdata == NULL)) {
123		return true;
124	}
125
126	/*
127	 * If this was the first creation of tdata, then
128	 * prof_tdata_get() reset bytes_until_sample, so decrement and
129	 * check it again
130	 */
131	if (!booted && prof_sample_check(tsd, usize, update)) {
132		return true;
133	}
134
135	if (tsd_reentrancy_level_get(tsd) > 0) {
136		return true;
137	}
138	/* Compute new sample threshold. */
139	if (update) {
140		prof_sample_threshold_update(tdata);
141	}
142	return !tdata->active;
143}
144
145JEMALLOC_ALWAYS_INLINE prof_tctx_t *
146prof_alloc_prep(tsd_t *tsd, size_t usize, bool prof_active, bool update) {
147	prof_tctx_t *ret;
148	prof_tdata_t *tdata;
149	prof_bt_t bt;
150
151	assert(usize == sz_s2u(usize));
152
153	if (!prof_active || likely(prof_sample_accum_update(tsd, usize, update,
154	    &tdata))) {
155		ret = (prof_tctx_t *)(uintptr_t)1U;
156	} else {
157		bt_init(&bt, tdata->vec);
158		prof_backtrace(&bt);
159		ret = prof_lookup(tsd, &bt);
160	}
161
162	return ret;
163}
164
165JEMALLOC_ALWAYS_INLINE void
166prof_malloc(tsdn_t *tsdn, const void *ptr, size_t usize, alloc_ctx_t *alloc_ctx,
167    prof_tctx_t *tctx) {
168	cassert(config_prof);
169	assert(ptr != NULL);
170	assert(usize == isalloc(tsdn, ptr));
171
172	if (unlikely((uintptr_t)tctx > (uintptr_t)1U)) {
173		prof_malloc_sample_object(tsdn, ptr, usize, tctx);
174	} else {
175		prof_tctx_set(tsdn, ptr, usize, alloc_ctx,
176		    (prof_tctx_t *)(uintptr_t)1U);
177	}
178}
179
180JEMALLOC_ALWAYS_INLINE void
181prof_realloc(tsd_t *tsd, const void *ptr, size_t usize, prof_tctx_t *tctx,
182    bool prof_active, bool updated, const void *old_ptr, size_t old_usize,
183    prof_tctx_t *old_tctx) {
184	bool sampled, old_sampled, moved;
185
186	cassert(config_prof);
187	assert(ptr != NULL || (uintptr_t)tctx <= (uintptr_t)1U);
188
189	if (prof_active && !updated && ptr != NULL) {
190		assert(usize == isalloc(tsd_tsdn(tsd), ptr));
191		if (prof_sample_accum_update(tsd, usize, true, NULL)) {
192			/*
193			 * Don't sample.  The usize passed to prof_alloc_prep()
194			 * was larger than what actually got allocated, so a
195			 * backtrace was captured for this allocation, even
196			 * though its actual usize was insufficient to cross the
197			 * sample threshold.
198			 */
199			prof_alloc_rollback(tsd, tctx, true);
200			tctx = (prof_tctx_t *)(uintptr_t)1U;
201		}
202	}
203
204	sampled = ((uintptr_t)tctx > (uintptr_t)1U);
205	old_sampled = ((uintptr_t)old_tctx > (uintptr_t)1U);
206	moved = (ptr != old_ptr);
207
208	if (unlikely(sampled)) {
209		prof_malloc_sample_object(tsd_tsdn(tsd), ptr, usize, tctx);
210	} else if (moved) {
211		prof_tctx_set(tsd_tsdn(tsd), ptr, usize, NULL,
212		    (prof_tctx_t *)(uintptr_t)1U);
213	} else if (unlikely(old_sampled)) {
214		/*
215		 * prof_tctx_set() would work for the !moved case as well, but
216		 * prof_tctx_reset() is slightly cheaper, and the proper thing
217		 * to do here in the presence of explicit knowledge re: moved
218		 * state.
219		 */
220		prof_tctx_reset(tsd_tsdn(tsd), ptr, tctx);
221	} else {
222		assert((uintptr_t)prof_tctx_get(tsd_tsdn(tsd), ptr, NULL) ==
223		    (uintptr_t)1U);
224	}
225
226	/*
227	 * The prof_free_sampled_object() call must come after the
228	 * prof_malloc_sample_object() call, because tctx and old_tctx may be
229	 * the same, in which case reversing the call order could cause the tctx
230	 * to be prematurely destroyed as a side effect of momentarily zeroed
231	 * counters.
232	 */
233	if (unlikely(old_sampled)) {
234		prof_free_sampled_object(tsd, ptr, old_usize, old_tctx);
235	}
236}
237
238JEMALLOC_ALWAYS_INLINE void
239prof_free(tsd_t *tsd, const void *ptr, size_t usize, alloc_ctx_t *alloc_ctx) {
240	prof_tctx_t *tctx = prof_tctx_get(tsd_tsdn(tsd), ptr, alloc_ctx);
241
242	cassert(config_prof);
243	assert(usize == isalloc(tsd_tsdn(tsd), ptr));
244
245	if (unlikely((uintptr_t)tctx > (uintptr_t)1U)) {
246		prof_free_sampled_object(tsd, ptr, usize, tctx);
247	}
248}
249
250#endif /* JEMALLOC_INTERNAL_PROF_INLINES_B_H */
251