|
1 /* |
|
2 Default header file for malloc-2.8.x, written by Doug Lea |
|
3 and released to the public domain, as explained at |
|
4 http://creativecommons.org/licenses/publicdomain. |
|
5 |
|
6 last update: Mon Aug 15 08:55:52 2005 Doug Lea (dl at gee) |
|
7 |
|
8 This header is for ANSI C/C++ only. You can set any of |
|
9 the following #defines before including: |
|
10 |
|
11 * If USE_DL_PREFIX is defined, it is assumed that malloc.c |
|
12 was also compiled with this option, so all routines |
|
13 have names starting with "dl". |
|
14 |
|
15 * If HAVE_USR_INCLUDE_MALLOC_H is defined, it is assumed that this |
|
16 file will be #included AFTER <malloc.h>. This is needed only if |
|
17 your system defines a struct mallinfo that is incompatible with the |
|
18 standard one declared here. Otherwise, you can include this file |
|
19 INSTEAD of your system system <malloc.h>. At least on ANSI, all |
|
20 declarations should be compatible with system versions |
|
21 |
|
22 * If MSPACES is defined, declarations for mspace versions are included. |
|
23 */ |
|
24 |
|
25 #ifndef MALLOC_280_H |
|
26 #define MALLOC_280_H |
|
27 |
|
28 #ifdef __cplusplus |
|
29 extern "C" { |
|
30 #endif |
|
31 |
|
32 #include <stddef.h> /* for size_t */ |
|
33 |
|
34 #if !ONLY_MSPACES |
|
35 |
|
36 #ifndef USE_DL_PREFIX |
|
37 #define dlcalloc calloc |
|
38 #define dlfree free |
|
39 #define dlmalloc malloc |
|
40 #define dlmemalign memalign |
|
41 #define dlrealloc realloc |
|
42 #define dlvalloc valloc |
|
43 #define dlpvalloc pvalloc |
|
44 #define dlmallinfo mallinfo |
|
45 #define dlmallopt mallopt |
|
46 #define dlmalloc_trim malloc_trim |
|
47 #define dlmalloc_stats malloc_stats |
|
48 #define dlmalloc_usable_size malloc_usable_size |
|
49 #define dlmalloc_footprint malloc_footprint |
|
50 #define dlindependent_calloc independent_calloc |
|
51 #define dlindependent_comalloc independent_comalloc |
|
52 #endif /* USE_DL_PREFIX */ |
|
53 |
|
54 |
|
55 /* |
|
56 malloc(size_t n) |
|
57 Returns a pointer to a newly allocated chunk of at least n bytes, or |
|
58 null if no space is available, in which case errno is set to ENOMEM |
|
59 on ANSI C systems. |
|
60 |
|
61 If n is zero, malloc returns a minimum-sized chunk. (The minimum |
|
62 size is 16 bytes on most 32bit systems, and 32 bytes on 64bit |
|
63 systems.) Note that size_t is an unsigned type, so calls with |
|
64 arguments that would be negative if signed are interpreted as |
|
65 requests for huge amounts of space, which will often fail. The |
|
66 maximum supported value of n differs across systems, but is in all |
|
67 cases less than the maximum representable value of a size_t. |
|
68 */ |
|
69 void* dlmalloc(size_t); |
|
70 |
|
71 /* |
|
72 free(void* p) |
|
73 Releases the chunk of memory pointed to by p, that had been previously |
|
74 allocated using malloc or a related routine such as realloc. |
|
75 It has no effect if p is null. If p was not malloced or already |
|
76 freed, free(p) will by default cuase the current program to abort. |
|
77 */ |
|
78 void dlfree(void*); |
|
79 |
|
80 /* |
|
81 calloc(size_t n_elements, size_t element_size); |
|
82 Returns a pointer to n_elements * element_size bytes, with all locations |
|
83 set to zero. |
|
84 */ |
|
85 void* dlcalloc(size_t, size_t); |
|
86 |
|
87 /* |
|
88 realloc(void* p, size_t n) |
|
89 Returns a pointer to a chunk of size n that contains the same data |
|
90 as does chunk p up to the minimum of (n, p's size) bytes, or null |
|
91 if no space is available. |
|
92 |
|
93 The returned pointer may or may not be the same as p. The algorithm |
|
94 prefers extending p in most cases when possible, otherwise it |
|
95 employs the equivalent of a malloc-copy-free sequence. |
|
96 |
|
97 If p is null, realloc is equivalent to malloc. |
|
98 |
|
99 If space is not available, realloc returns null, errno is set (if on |
|
100 ANSI) and p is NOT freed. |
|
101 |
|
102 if n is for fewer bytes than already held by p, the newly unused |
|
103 space is lopped off and freed if possible. realloc with a size |
|
104 argument of zero (re)allocates a minimum-sized chunk. |
|
105 |
|
106 The old unix realloc convention of allowing the last-free'd chunk |
|
107 to be used as an argument to realloc is not supported. |
|
108 */ |
|
109 |
|
110 void* dlrealloc(void*, size_t); |
|
111 |
|
112 /* |
|
113 memalign(size_t alignment, size_t n); |
|
114 Returns a pointer to a newly allocated chunk of n bytes, aligned |
|
115 in accord with the alignment argument. |
|
116 |
|
117 The alignment argument should be a power of two. If the argument is |
|
118 not a power of two, the nearest greater power is used. |
|
119 8-byte alignment is guaranteed by normal malloc calls, so don't |
|
120 bother calling memalign with an argument of 8 or less. |
|
121 |
|
122 Overreliance on memalign is a sure way to fragment space. |
|
123 */ |
|
124 void* dlmemalign(size_t, size_t); |
|
125 |
|
126 /* |
|
127 valloc(size_t n); |
|
128 Equivalent to memalign(pagesize, n), where pagesize is the page |
|
129 size of the system. If the pagesize is unknown, 4096 is used. |
|
130 */ |
|
131 void* dlvalloc(size_t); |
|
132 |
|
133 /* |
|
134 mallopt(int parameter_number, int parameter_value) |
|
135 Sets tunable parameters The format is to provide a |
|
136 (parameter-number, parameter-value) pair. mallopt then sets the |
|
137 corresponding parameter to the argument value if it can (i.e., so |
|
138 long as the value is meaningful), and returns 1 if successful else |
|
139 0. SVID/XPG/ANSI defines four standard param numbers for mallopt, |
|
140 normally defined in malloc.h. None of these are use in this malloc, |
|
141 so setting them has no effect. But this malloc also supports other |
|
142 options in mallopt: |
|
143 |
|
144 Symbol param # default allowed param values |
|
145 M_TRIM_THRESHOLD -1 2*1024*1024 any (-1U disables trimming) |
|
146 M_GRANULARITY -2 page size any power of 2 >= page size |
|
147 M_MMAP_THRESHOLD -3 256*1024 any (or 0 if no MMAP support) |
|
148 */ |
|
149 int dlmallopt(int, int); |
|
150 |
|
151 #define M_TRIM_THRESHOLD (-1) |
|
152 #define M_GRANULARITY (-2) |
|
153 #define M_MMAP_THRESHOLD (-3) |
|
154 |
|
155 |
|
156 /* |
|
157 malloc_footprint(); |
|
158 Returns the number of bytes obtained from the system. The total |
|
159 number of bytes allocated by malloc, realloc etc., is less than this |
|
160 value. Unlike mallinfo, this function returns only a precomputed |
|
161 result, so can be called frequently to monitor memory consumption. |
|
162 Even if locks are otherwise defined, this function does not use them, |
|
163 so results might not be up to date. |
|
164 */ |
|
165 size_t dlmalloc_footprint(void); |
|
166 |
|
167 #if !NO_MALLINFO |
|
168 /* |
|
169 mallinfo() |
|
170 Returns (by copy) a struct containing various summary statistics: |
|
171 |
|
172 arena: current total non-mmapped bytes allocated from system |
|
173 ordblks: the number of free chunks |
|
174 smblks: always zero. |
|
175 hblks: current number of mmapped regions |
|
176 hblkhd: total bytes held in mmapped regions |
|
177 usmblks: the maximum total allocated space. This will be greater |
|
178 than current total if trimming has occurred. |
|
179 fsmblks: always zero |
|
180 uordblks: current total allocated space (normal or mmapped) |
|
181 fordblks: total free space |
|
182 keepcost: the maximum number of bytes that could ideally be released |
|
183 back to system via malloc_trim. ("ideally" means that |
|
184 it ignores page restrictions etc.) |
|
185 |
|
186 Because these fields are ints, but internal bookkeeping may |
|
187 be kept as longs, the reported values may wrap around zero and |
|
188 thus be inaccurate. |
|
189 */ |
|
190 #ifndef HAVE_USR_INCLUDE_MALLOC_H |
|
191 #ifndef _MALLOC_H |
|
192 #ifndef MALLINFO_FIELD_TYPE |
|
193 #define MALLINFO_FIELD_TYPE size_t |
|
194 #endif /* MALLINFO_FIELD_TYPE */ |
|
195 struct mallinfo { |
|
196 MALLINFO_FIELD_TYPE arena; /* non-mmapped space allocated from system */ |
|
197 MALLINFO_FIELD_TYPE ordblks; /* number of free chunks */ |
|
198 MALLINFO_FIELD_TYPE smblks; /* always 0 */ |
|
199 MALLINFO_FIELD_TYPE hblks; /* always 0 */ |
|
200 MALLINFO_FIELD_TYPE hblkhd; /* space in mmapped regions */ |
|
201 MALLINFO_FIELD_TYPE usmblks; /* maximum total allocated space */ |
|
202 MALLINFO_FIELD_TYPE fsmblks; /* always 0 */ |
|
203 MALLINFO_FIELD_TYPE uordblks; /* total allocated space */ |
|
204 MALLINFO_FIELD_TYPE fordblks; /* total free space */ |
|
205 MALLINFO_FIELD_TYPE keepcost; /* releasable (via malloc_trim) space */ |
|
206 }; |
|
207 #endif /* _MALLOC_H */ |
|
208 #endif /* HAVE_USR_INCLUDE_MALLOC_H */ |
|
209 |
|
210 struct mallinfo dlmallinfo(void); |
|
211 #endif /* NO_MALLINFO */ |
|
212 |
|
213 /* |
|
214 independent_calloc(size_t n_elements, size_t element_size, void* chunks[]); |
|
215 |
|
216 independent_calloc is similar to calloc, but instead of returning a |
|
217 single cleared space, it returns an array of pointers to n_elements |
|
218 independent elements that can hold contents of size elem_size, each |
|
219 of which starts out cleared, and can be independently freed, |
|
220 realloc'ed etc. The elements are guaranteed to be adjacently |
|
221 allocated (this is not guaranteed to occur with multiple callocs or |
|
222 mallocs), which may also improve cache locality in some |
|
223 applications. |
|
224 |
|
225 The "chunks" argument is optional (i.e., may be null, which is |
|
226 probably the most typical usage). If it is null, the returned array |
|
227 is itself dynamically allocated and should also be freed when it is |
|
228 no longer needed. Otherwise, the chunks array must be of at least |
|
229 n_elements in length. It is filled in with the pointers to the |
|
230 chunks. |
|
231 |
|
232 In either case, independent_calloc returns this pointer array, or |
|
233 null if the allocation failed. If n_elements is zero and "chunks" |
|
234 is null, it returns a chunk representing an array with zero elements |
|
235 (which should be freed if not wanted). |
|
236 |
|
237 Each element must be individually freed when it is no longer |
|
238 needed. If you'd like to instead be able to free all at once, you |
|
239 should instead use regular calloc and assign pointers into this |
|
240 space to represent elements. (In this case though, you cannot |
|
241 independently free elements.) |
|
242 |
|
243 independent_calloc simplifies and speeds up implementations of many |
|
244 kinds of pools. It may also be useful when constructing large data |
|
245 structures that initially have a fixed number of fixed-sized nodes, |
|
246 but the number is not known at compile time, and some of the nodes |
|
247 may later need to be freed. For example: |
|
248 |
|
249 struct Node { int item; struct Node* next; }; |
|
250 |
|
251 struct Node* build_list() { |
|
252 struct Node** pool; |
|
253 int n = read_number_of_nodes_needed(); |
|
254 if (n <= 0) return 0; |
|
255 pool = (struct Node**)(independent_calloc(n, sizeof(struct Node), 0); |
|
256 if (pool == 0) die(); |
|
257 // organize into a linked list... |
|
258 struct Node* first = pool[0]; |
|
259 for (i = 0; i < n-1; ++i) |
|
260 pool[i]->next = pool[i+1]; |
|
261 free(pool); // Can now free the array (or not, if it is needed later) |
|
262 return first; |
|
263 } |
|
264 */ |
|
265 void** dlindependent_calloc(size_t, size_t, void**); |
|
266 |
|
267 /* |
|
268 independent_comalloc(size_t n_elements, size_t sizes[], void* chunks[]); |
|
269 |
|
270 independent_comalloc allocates, all at once, a set of n_elements |
|
271 chunks with sizes indicated in the "sizes" array. It returns |
|
272 an array of pointers to these elements, each of which can be |
|
273 independently freed, realloc'ed etc. The elements are guaranteed to |
|
274 be adjacently allocated (this is not guaranteed to occur with |
|
275 multiple callocs or mallocs), which may also improve cache locality |
|
276 in some applications. |
|
277 |
|
278 The "chunks" argument is optional (i.e., may be null). If it is null |
|
279 the returned array is itself dynamically allocated and should also |
|
280 be freed when it is no longer needed. Otherwise, the chunks array |
|
281 must be of at least n_elements in length. It is filled in with the |
|
282 pointers to the chunks. |
|
283 |
|
284 In either case, independent_comalloc returns this pointer array, or |
|
285 null if the allocation failed. If n_elements is zero and chunks is |
|
286 null, it returns a chunk representing an array with zero elements |
|
287 (which should be freed if not wanted). |
|
288 |
|
289 Each element must be individually freed when it is no longer |
|
290 needed. If you'd like to instead be able to free all at once, you |
|
291 should instead use a single regular malloc, and assign pointers at |
|
292 particular offsets in the aggregate space. (In this case though, you |
|
293 cannot independently free elements.) |
|
294 |
|
295 independent_comallac differs from independent_calloc in that each |
|
296 element may have a different size, and also that it does not |
|
297 automatically clear elements. |
|
298 |
|
299 independent_comalloc can be used to speed up allocation in cases |
|
300 where several structs or objects must always be allocated at the |
|
301 same time. For example: |
|
302 |
|
303 struct Head { ... } |
|
304 struct Foot { ... } |
|
305 |
|
306 void send_message(char* msg) { |
|
307 int msglen = strlen(msg); |
|
308 size_t sizes[3] = { sizeof(struct Head), msglen, sizeof(struct Foot) }; |
|
309 void* chunks[3]; |
|
310 if (independent_comalloc(3, sizes, chunks) == 0) |
|
311 die(); |
|
312 struct Head* head = (struct Head*)(chunks[0]); |
|
313 char* body = (char*)(chunks[1]); |
|
314 struct Foot* foot = (struct Foot*)(chunks[2]); |
|
315 // ... |
|
316 } |
|
317 |
|
318 In general though, independent_comalloc is worth using only for |
|
319 larger values of n_elements. For small values, you probably won't |
|
320 detect enough difference from series of malloc calls to bother. |
|
321 |
|
322 Overuse of independent_comalloc can increase overall memory usage, |
|
323 since it cannot reuse existing noncontiguous small chunks that |
|
324 might be available for some of the elements. |
|
325 */ |
|
326 void** dlindependent_comalloc(size_t, size_t*, void**); |
|
327 |
|
328 |
|
329 /* |
|
330 pvalloc(size_t n); |
|
331 Equivalent to valloc(minimum-page-that-holds(n)), that is, |
|
332 round up n to nearest pagesize. |
|
333 */ |
|
334 void* dlpvalloc(size_t); |
|
335 |
|
336 /* |
|
337 malloc_trim(size_t pad); |
|
338 |
|
339 If possible, gives memory back to the system (via negative arguments |
|
340 to sbrk) if there is unused memory at the `high' end of the malloc |
|
341 pool or in unused MMAP segments. You can call this after freeing |
|
342 large blocks of memory to potentially reduce the system-level memory |
|
343 requirements of a program. However, it cannot guarantee to reduce |
|
344 memory. Under some allocation patterns, some large free blocks of |
|
345 memory will be locked between two used chunks, so they cannot be |
|
346 given back to the system. |
|
347 |
|
348 The `pad' argument to malloc_trim represents the amount of free |
|
349 trailing space to leave untrimmed. If this argument is zero, only |
|
350 the minimum amount of memory to maintain internal data structures |
|
351 will be left. Non-zero arguments can be supplied to maintain enough |
|
352 trailing space to service future expected allocations without having |
|
353 to re-obtain memory from the system. |
|
354 |
|
355 Malloc_trim returns 1 if it actually released any memory, else 0. |
|
356 */ |
|
357 int dlmalloc_trim(size_t); |
|
358 |
|
359 /* |
|
360 malloc_usable_size(void* p); |
|
361 |
|
362 Returns the number of bytes you can actually use in |
|
363 an allocated chunk, which may be more than you requested (although |
|
364 often not) due to alignment and minimum size constraints. |
|
365 You can use this many bytes without worrying about |
|
366 overwriting other allocated objects. This is not a particularly great |
|
367 programming practice. malloc_usable_size can be more useful in |
|
368 debugging and assertions, for example: |
|
369 |
|
370 p = malloc(n); |
|
371 assert(malloc_usable_size(p) >= 256); |
|
372 */ |
|
373 size_t dlmalloc_usable_size(void*); |
|
374 |
|
375 /* |
|
376 malloc_stats(); |
|
377 Prints on stderr the amount of space obtained from the system (both |
|
378 via sbrk and mmap), the maximum amount (which may be more than |
|
379 current if malloc_trim and/or munmap got called), and the current |
|
380 number of bytes allocated via malloc (or realloc, etc) but not yet |
|
381 freed. Note that this is the number of bytes allocated, not the |
|
382 number requested. It will be larger than the number requested |
|
383 because of alignment and bookkeeping overhead. Because it includes |
|
384 alignment wastage as being in use, this figure may be greater than |
|
385 zero even when no user-level chunks are allocated. |
|
386 |
|
387 The reported current and maximum system memory can be inaccurate if |
|
388 a program makes other calls to system memory allocation functions |
|
389 (normally sbrk) outside of malloc. |
|
390 |
|
391 malloc_stats prints only the most commonly interesting statistics. |
|
392 More information can be obtained by calling mallinfo. |
|
393 */ |
|
394 void dlmalloc_stats(void); |
|
395 |
|
396 #endif /* !ONLY_MSPACES */ |
|
397 |
|
398 #if MSPACES |
|
399 |
|
400 /* |
|
401 mspace is an opaque type representing an independent |
|
402 region of space that supports mspace_malloc, etc. |
|
403 */ |
|
404 typedef void* mspace; |
|
405 |
|
406 /* |
|
407 create_mspace creates and returns a new independent space with the |
|
408 given initial capacity, or, if 0, the default granularity size. It |
|
409 returns null if there is no system memory available to create the |
|
410 space. If argument locked is non-zero, the space uses a separate |
|
411 lock to control access. The capacity of the space will grow |
|
412 dynamically as needed to service mspace_malloc requests. You can |
|
413 control the sizes of incremental increases of this space by |
|
414 compiling with a different DEFAULT_GRANULARITY or dynamically |
|
415 setting with mallopt(M_GRANULARITY, value). |
|
416 */ |
|
417 mspace create_mspace(size_t capacity, int locked); |
|
418 |
|
419 /* |
|
420 destroy_mspace destroys the given space, and attempts to return all |
|
421 of its memory back to the system, returning the total number of |
|
422 bytes freed. After destruction, the results of access to all memory |
|
423 used by the space become undefined. |
|
424 */ |
|
425 size_t destroy_mspace(mspace msp); |
|
426 |
|
427 /* |
|
428 create_mspace_with_base uses the memory supplied as the initial base |
|
429 of a new mspace. Part (less than 128*sizeof(size_t) bytes) of this |
|
430 space is used for bookkeeping, so the capacity must be at least this |
|
431 large. (Otherwise 0 is returned.) When this initial space is |
|
432 exhausted, additional memory will be obtained from the system. |
|
433 Destroying this space will deallocate all additionally allocated |
|
434 space (if possible) but not the initial base. |
|
435 */ |
|
436 mspace create_mspace_with_base(void* base, size_t capacity, int locked); |
|
437 |
|
438 /* |
|
439 mspace_malloc behaves as malloc, but operates within |
|
440 the given space. |
|
441 */ |
|
442 void* mspace_malloc(mspace msp, size_t bytes); |
|
443 |
|
444 /* |
|
445 mspace_free behaves as free, but operates within |
|
446 the given space. |
|
447 |
|
448 If compiled with FOOTERS==1, mspace_free is not actually needed. |
|
449 free may be called instead of mspace_free because freed chunks from |
|
450 any space are handled by their originating spaces. |
|
451 */ |
|
452 void mspace_free(mspace msp, void* mem); |
|
453 |
|
454 /* |
|
455 mspace_realloc behaves as realloc, but operates within |
|
456 the given space. |
|
457 |
|
458 If compiled with FOOTERS==1, mspace_realloc is not actually |
|
459 needed. realloc may be called instead of mspace_realloc because |
|
460 realloced chunks from any space are handled by their originating |
|
461 spaces. |
|
462 */ |
|
463 void* mspace_realloc(mspace msp, void* mem, size_t newsize); |
|
464 |
|
465 /* |
|
466 mspace_calloc behaves as calloc, but operates within |
|
467 the given space. |
|
468 */ |
|
469 void* mspace_calloc(mspace msp, size_t n_elements, size_t elem_size); |
|
470 |
|
471 /* |
|
472 mspace_memalign behaves as memalign, but operates within |
|
473 the given space. |
|
474 */ |
|
475 void* mspace_memalign(mspace msp, size_t alignment, size_t bytes); |
|
476 |
|
477 /* |
|
478 mspace_independent_calloc behaves as independent_calloc, but |
|
479 operates within the given space. |
|
480 */ |
|
481 void** mspace_independent_calloc(mspace msp, size_t n_elements, |
|
482 size_t elem_size, void* chunks[]); |
|
483 |
|
484 /* |
|
485 mspace_independent_comalloc behaves as independent_comalloc, but |
|
486 operates within the given space. |
|
487 */ |
|
488 void** mspace_independent_comalloc(mspace msp, size_t n_elements, |
|
489 size_t sizes[], void* chunks[]); |
|
490 |
|
491 /* |
|
492 mspace_footprint() returns the number of bytes obtained from the |
|
493 system for this space. |
|
494 */ |
|
495 size_t mspace_footprint(mspace msp); |
|
496 |
|
497 |
|
498 #if !NO_MALLINFO |
|
499 /* |
|
500 mspace_mallinfo behaves as mallinfo, but reports properties of |
|
501 the given space. |
|
502 */ |
|
503 struct mallinfo mspace_mallinfo(mspace msp); |
|
504 #endif /* NO_MALLINFO */ |
|
505 |
|
506 /* |
|
507 mspace_malloc_stats behaves as malloc_stats, but reports |
|
508 properties of the given space. |
|
509 */ |
|
510 void mspace_malloc_stats(mspace msp); |
|
511 |
|
512 /* |
|
513 mspace_trim behaves as malloc_trim, but |
|
514 operates within the given space. |
|
515 */ |
|
516 int mspace_trim(mspace msp, size_t pad); |
|
517 |
|
518 /* |
|
519 An alias for malloc_usable_size. |
|
520 */ |
|
521 size_t mspace_usable_size(void *mem); |
|
522 |
|
523 /* |
|
524 An alias for mallopt. |
|
525 */ |
|
526 int mspace_mallopt(int, int); |
|
527 |
|
528 #endif /* MSPACES */ |
|
529 |
|
530 #ifdef __cplusplus |
|
531 }; /* end of extern "C" */ |
|
532 #endif |
|
533 |
|
534 #endif /* MALLOC_280_H */ |