]> git.ipfire.org Git - people/ms/u-boot.git/blame - include/malloc.h
* Patches by Robert Schwebel, 26 Jun 2003:
[people/ms/u-boot.git] / include / malloc.h
CommitLineData
5b1d7137
WD
1/*
2 A version of malloc/free/realloc written by Doug Lea and released to the
3 public domain. Send questions/comments/complaints/performance data
4 to dl@cs.oswego.edu
5
6* VERSION 2.6.6 Sun Mar 5 19:10:03 2000 Doug Lea (dl at gee)
7
8 Note: There may be an updated version of this malloc obtainable at
9 ftp://g.oswego.edu/pub/misc/malloc.c
10 Check before installing!
11
12* Why use this malloc?
13
14 This is not the fastest, most space-conserving, most portable, or
15 most tunable malloc ever written. However it is among the fastest
16 while also being among the most space-conserving, portable and tunable.
17 Consistent balance across these factors results in a good general-purpose
18 allocator. For a high-level description, see
19 http://g.oswego.edu/dl/html/malloc.html
20
21* Synopsis of public routines
22
23 (Much fuller descriptions are contained in the program documentation below.)
24
25 malloc(size_t n);
26 Return a pointer to a newly allocated chunk of at least n bytes, or null
27 if no space is available.
28 free(Void_t* p);
29 Release the chunk of memory pointed to by p, or no effect if p is null.
30 realloc(Void_t* p, size_t n);
31 Return a pointer to a chunk of size n that contains the same data
32 as does chunk p up to the minimum of (n, p's size) bytes, or null
33 if no space is available. The returned pointer may or may not be
34 the same as p. If p is null, equivalent to malloc. Unless the
35 #define REALLOC_ZERO_BYTES_FREES below is set, realloc with a
36 size argument of zero (re)allocates a minimum-sized chunk.
37 memalign(size_t alignment, size_t n);
38 Return a pointer to a newly allocated chunk of n bytes, aligned
39 in accord with the alignment argument, which must be a power of
40 two.
41 valloc(size_t n);
42 Equivalent to memalign(pagesize, n), where pagesize is the page
43 size of the system (or as near to this as can be figured out from
44 all the includes/defines below.)
45 pvalloc(size_t n);
46 Equivalent to valloc(minimum-page-that-holds(n)), that is,
47 round up n to nearest pagesize.
48 calloc(size_t unit, size_t quantity);
49 Returns a pointer to quantity * unit bytes, with all locations
50 set to zero.
51 cfree(Void_t* p);
52 Equivalent to free(p).
53 malloc_trim(size_t pad);
54 Release all but pad bytes of freed top-most memory back
55 to the system. Return 1 if successful, else 0.
56 malloc_usable_size(Void_t* p);
57 Report the number usable allocated bytes associated with allocated
58 chunk p. This may or may not report more bytes than were requested,
59 due to alignment and minimum size constraints.
60 malloc_stats();
61 Prints brief summary statistics on stderr.
62 mallinfo()
63 Returns (by copy) a struct containing various summary statistics.
64 mallopt(int parameter_number, int parameter_value)
65 Changes one of the tunable parameters described below. Returns
66 1 if successful in changing the parameter, else 0.
67
68* Vital statistics:
69
70 Alignment: 8-byte
71 8 byte alignment is currently hardwired into the design. This
72 seems to suffice for all current machines and C compilers.
73
74 Assumed pointer representation: 4 or 8 bytes
75 Code for 8-byte pointers is untested by me but has worked
76 reliably by Wolfram Gloger, who contributed most of the
77 changes supporting this.
78
79 Assumed size_t representation: 4 or 8 bytes
80 Note that size_t is allowed to be 4 bytes even if pointers are 8.
81
82 Minimum overhead per allocated chunk: 4 or 8 bytes
83 Each malloced chunk has a hidden overhead of 4 bytes holding size
84 and status information.
85
86 Minimum allocated size: 4-byte ptrs: 16 bytes (including 4 overhead)
87 8-byte ptrs: 24/32 bytes (including, 4/8 overhead)
88
89 When a chunk is freed, 12 (for 4byte ptrs) or 20 (for 8 byte
90 ptrs but 4 byte size) or 24 (for 8/8) additional bytes are
91 needed; 4 (8) for a trailing size field
92 and 8 (16) bytes for free list pointers. Thus, the minimum
93 allocatable size is 16/24/32 bytes.
94
95 Even a request for zero bytes (i.e., malloc(0)) returns a
96 pointer to something of the minimum allocatable size.
97
98 Maximum allocated size: 4-byte size_t: 2^31 - 8 bytes
99 8-byte size_t: 2^63 - 16 bytes
100
101 It is assumed that (possibly signed) size_t bit values suffice to
102 represent chunk sizes. `Possibly signed' is due to the fact
103 that `size_t' may be defined on a system as either a signed or
104 an unsigned type. To be conservative, values that would appear
105 as negative numbers are avoided.
106 Requests for sizes with a negative sign bit when the request
107 size is treaded as a long will return null.
108
109 Maximum overhead wastage per allocated chunk: normally 15 bytes
110
111 Alignnment demands, plus the minimum allocatable size restriction
112 make the normal worst-case wastage 15 bytes (i.e., up to 15
113 more bytes will be allocated than were requested in malloc), with
114 two exceptions:
115 1. Because requests for zero bytes allocate non-zero space,
116 the worst case wastage for a request of zero bytes is 24 bytes.
117 2. For requests >= mmap_threshold that are serviced via
118 mmap(), the worst case wastage is 8 bytes plus the remainder
119 from a system page (the minimal mmap unit); typically 4096 bytes.
120
121* Limitations
122
123 Here are some features that are NOT currently supported
124
125 * No user-definable hooks for callbacks and the like.
126 * No automated mechanism for fully checking that all accesses
127 to malloced memory stay within their bounds.
128 * No support for compaction.
129
130* Synopsis of compile-time options:
131
132 People have reported using previous versions of this malloc on all
133 versions of Unix, sometimes by tweaking some of the defines
134 below. It has been tested most extensively on Solaris and
135 Linux. It is also reported to work on WIN32 platforms.
136 People have also reported adapting this malloc for use in
137 stand-alone embedded systems.
138
139 The implementation is in straight, hand-tuned ANSI C. Among other
140 consequences, it uses a lot of macros. Because of this, to be at
141 all usable, this code should be compiled using an optimizing compiler
142 (for example gcc -O2) that can simplify expressions and control
143 paths.
144
145 __STD_C (default: derived from C compiler defines)
146 Nonzero if using ANSI-standard C compiler, a C++ compiler, or
147 a C compiler sufficiently close to ANSI to get away with it.
148 DEBUG (default: NOT defined)
149 Define to enable debugging. Adds fairly extensive assertion-based
150 checking to help track down memory errors, but noticeably slows down
151 execution.
152 REALLOC_ZERO_BYTES_FREES (default: NOT defined)
153 Define this if you think that realloc(p, 0) should be equivalent
154 to free(p). Otherwise, since malloc returns a unique pointer for
155 malloc(0), so does realloc(p, 0).
156 HAVE_MEMCPY (default: defined)
157 Define if you are not otherwise using ANSI STD C, but still
158 have memcpy and memset in your C library and want to use them.
159 Otherwise, simple internal versions are supplied.
160 USE_MEMCPY (default: 1 if HAVE_MEMCPY is defined, 0 otherwise)
161 Define as 1 if you want the C library versions of memset and
162 memcpy called in realloc and calloc (otherwise macro versions are used).
163 At least on some platforms, the simple macro versions usually
164 outperform libc versions.
165 HAVE_MMAP (default: defined as 1)
166 Define to non-zero to optionally make malloc() use mmap() to
167 allocate very large blocks.
168 HAVE_MREMAP (default: defined as 0 unless Linux libc set)
169 Define to non-zero to optionally make realloc() use mremap() to
170 reallocate very large blocks.
171 malloc_getpagesize (default: derived from system #includes)
172 Either a constant or routine call returning the system page size.
173 HAVE_USR_INCLUDE_MALLOC_H (default: NOT defined)
174 Optionally define if you are on a system with a /usr/include/malloc.h
175 that declares struct mallinfo. It is not at all necessary to
176 define this even if you do, but will ensure consistency.
177 INTERNAL_SIZE_T (default: size_t)
178 Define to a 32-bit type (probably `unsigned int') if you are on a
179 64-bit machine, yet do not want or need to allow malloc requests of
180 greater than 2^31 to be handled. This saves space, especially for
181 very small chunks.
182 INTERNAL_LINUX_C_LIB (default: NOT defined)
183 Defined only when compiled as part of Linux libc.
184 Also note that there is some odd internal name-mangling via defines
185 (for example, internally, `malloc' is named `mALLOc') needed
186 when compiling in this case. These look funny but don't otherwise
187 affect anything.
188 WIN32 (default: undefined)
189 Define this on MS win (95, nt) platforms to compile in sbrk emulation.
190 LACKS_UNISTD_H (default: undefined if not WIN32)
191 Define this if your system does not have a <unistd.h>.
192 LACKS_SYS_PARAM_H (default: undefined if not WIN32)
193 Define this if your system does not have a <sys/param.h>.
194 MORECORE (default: sbrk)
195 The name of the routine to call to obtain more memory from the system.
196 MORECORE_FAILURE (default: -1)
197 The value returned upon failure of MORECORE.
198 MORECORE_CLEARS (default 1)
199 True (1) if the routine mapped to MORECORE zeroes out memory (which
200 holds for sbrk).
201 DEFAULT_TRIM_THRESHOLD
202 DEFAULT_TOP_PAD
203 DEFAULT_MMAP_THRESHOLD
204 DEFAULT_MMAP_MAX
205 Default values of tunable parameters (described in detail below)
206 controlling interaction with host system routines (sbrk, mmap, etc).
207 These values may also be changed dynamically via mallopt(). The
208 preset defaults are those that give best performance for typical
209 programs/systems.
210 USE_DL_PREFIX (default: undefined)
211 Prefix all public routines with the string 'dl'. Useful to
212 quickly avoid procedure declaration conflicts and linker symbol
213 conflicts with existing memory allocation routines.
214
215
216*/
217
218\f
219
220
221/* Preliminaries */
222
223#ifndef __STD_C
224#ifdef __STDC__
225#define __STD_C 1
226#else
227#if __cplusplus
228#define __STD_C 1
229#else
230#define __STD_C 0
231#endif /*__cplusplus*/
232#endif /*__STDC__*/
233#endif /*__STD_C*/
234
235#ifndef Void_t
236#if (__STD_C || defined(WIN32))
237#define Void_t void
238#else
239#define Void_t char
240#endif
241#endif /*Void_t*/
242
243#if __STD_C
244#include <linux/stddef.h> /* for size_t */
245#else
246#include <sys/types.h>
247#endif /* __STD_C */
248
249#ifdef __cplusplus
250extern "C" {
251#endif
252
253#if 0 /* not for U-Boot */
254#include <stdio.h> /* needed for malloc_stats */
255#endif
256
257
258/*
259 Compile-time options
260*/
261
262
263/*
264 Debugging:
265
266 Because freed chunks may be overwritten with link fields, this
267 malloc will often die when freed memory is overwritten by user
268 programs. This can be very effective (albeit in an annoying way)
269 in helping track down dangling pointers.
270
271 If you compile with -DDEBUG, a number of assertion checks are
272 enabled that will catch more memory errors. You probably won't be
273 able to make much sense of the actual assertion errors, but they
274 should help you locate incorrectly overwritten memory. The
275 checking is fairly extensive, and will slow down execution
276 noticeably. Calling malloc_stats or mallinfo with DEBUG set will
277 attempt to check every non-mmapped allocated and free chunk in the
278 course of computing the summmaries. (By nature, mmapped regions
279 cannot be checked very much automatically.)
280
281 Setting DEBUG may also be helpful if you are trying to modify
282 this code. The assertions in the check routines spell out in more
283 detail the assumptions and invariants underlying the algorithms.
284
285*/
286
287#ifdef DEBUG
288/* #include <assert.h> */
289#define assert(x) ((void)0)
290#else
291#define assert(x) ((void)0)
292#endif
293
294
295/*
296 INTERNAL_SIZE_T is the word-size used for internal bookkeeping
297 of chunk sizes. On a 64-bit machine, you can reduce malloc
298 overhead by defining INTERNAL_SIZE_T to be a 32 bit `unsigned int'
299 at the expense of not being able to handle requests greater than
300 2^31. This limitation is hardly ever a concern; you are encouraged
301 to set this. However, the default version is the same as size_t.
302*/
303
304#ifndef INTERNAL_SIZE_T
305#define INTERNAL_SIZE_T size_t
306#endif
307
308/*
309 REALLOC_ZERO_BYTES_FREES should be set if a call to
310 realloc with zero bytes should be the same as a call to free.
311 Some people think it should. Otherwise, since this malloc
312 returns a unique pointer for malloc(0), so does realloc(p, 0).
313*/
314
315
316/* #define REALLOC_ZERO_BYTES_FREES */
317
318
319/*
320 WIN32 causes an emulation of sbrk to be compiled in
321 mmap-based options are not currently supported in WIN32.
322*/
323
324/* #define WIN32 */
325#ifdef WIN32
326#define MORECORE wsbrk
327#define HAVE_MMAP 0
328
329#define LACKS_UNISTD_H
330#define LACKS_SYS_PARAM_H
331
332/*
333 Include 'windows.h' to get the necessary declarations for the
334 Microsoft Visual C++ data structures and routines used in the 'sbrk'
335 emulation.
336
337 Define WIN32_LEAN_AND_MEAN so that only the essential Microsoft
338 Visual C++ header files are included.
339*/
340#define WIN32_LEAN_AND_MEAN
341#include <windows.h>
342#endif
343
344
345/*
346 HAVE_MEMCPY should be defined if you are not otherwise using
347 ANSI STD C, but still have memcpy and memset in your C library
348 and want to use them in calloc and realloc. Otherwise simple
349 macro versions are defined here.
350
351 USE_MEMCPY should be defined as 1 if you actually want to
352 have memset and memcpy called. People report that the macro
353 versions are often enough faster than libc versions on many
354 systems that it is better to use them.
355
356*/
357
358#define HAVE_MEMCPY
359
360#ifndef USE_MEMCPY
361#ifdef HAVE_MEMCPY
362#define USE_MEMCPY 1
363#else
364#define USE_MEMCPY 0
365#endif
366#endif
367
368#if (__STD_C || defined(HAVE_MEMCPY))
369
370#if __STD_C
371void* memset(void*, int, size_t);
372void* memcpy(void*, const void*, size_t);
373#else
374#ifdef WIN32
375// On Win32 platforms, 'memset()' and 'memcpy()' are already declared in
376// 'windows.h'
377#else
378Void_t* memset();
379Void_t* memcpy();
380#endif
381#endif
382#endif
383
384#if USE_MEMCPY
385
386/* The following macros are only invoked with (2n+1)-multiples of
387 INTERNAL_SIZE_T units, with a positive integer n. This is exploited
388 for fast inline execution when n is small. */
389
390#define MALLOC_ZERO(charp, nbytes) \
391do { \
392 INTERNAL_SIZE_T mzsz = (nbytes); \
393 if(mzsz <= 9*sizeof(mzsz)) { \
394 INTERNAL_SIZE_T* mz = (INTERNAL_SIZE_T*) (charp); \
395 if(mzsz >= 5*sizeof(mzsz)) { *mz++ = 0; \
396 *mz++ = 0; \
397 if(mzsz >= 7*sizeof(mzsz)) { *mz++ = 0; \
398 *mz++ = 0; \
399 if(mzsz >= 9*sizeof(mzsz)) { *mz++ = 0; \
400 *mz++ = 0; }}} \
401 *mz++ = 0; \
402 *mz++ = 0; \
403 *mz = 0; \
404 } else memset((charp), 0, mzsz); \
405} while(0)
406
407#define MALLOC_COPY(dest,src,nbytes) \
408do { \
409 INTERNAL_SIZE_T mcsz = (nbytes); \
410 if(mcsz <= 9*sizeof(mcsz)) { \
411 INTERNAL_SIZE_T* mcsrc = (INTERNAL_SIZE_T*) (src); \
412 INTERNAL_SIZE_T* mcdst = (INTERNAL_SIZE_T*) (dest); \
413 if(mcsz >= 5*sizeof(mcsz)) { *mcdst++ = *mcsrc++; \
414 *mcdst++ = *mcsrc++; \
415 if(mcsz >= 7*sizeof(mcsz)) { *mcdst++ = *mcsrc++; \
416 *mcdst++ = *mcsrc++; \
417 if(mcsz >= 9*sizeof(mcsz)) { *mcdst++ = *mcsrc++; \
418 *mcdst++ = *mcsrc++; }}} \
419 *mcdst++ = *mcsrc++; \
420 *mcdst++ = *mcsrc++; \
421 *mcdst = *mcsrc ; \
422 } else memcpy(dest, src, mcsz); \
423} while(0)
424
425#else /* !USE_MEMCPY */
426
427/* Use Duff's device for good zeroing/copying performance. */
428
429#define MALLOC_ZERO(charp, nbytes) \
430do { \
431 INTERNAL_SIZE_T* mzp = (INTERNAL_SIZE_T*)(charp); \
432 long mctmp = (nbytes)/sizeof(INTERNAL_SIZE_T), mcn; \
433 if (mctmp < 8) mcn = 0; else { mcn = (mctmp-1)/8; mctmp %= 8; } \
434 switch (mctmp) { \
435 case 0: for(;;) { *mzp++ = 0; \
436 case 7: *mzp++ = 0; \
437 case 6: *mzp++ = 0; \
438 case 5: *mzp++ = 0; \
439 case 4: *mzp++ = 0; \
440 case 3: *mzp++ = 0; \
441 case 2: *mzp++ = 0; \
442 case 1: *mzp++ = 0; if(mcn <= 0) break; mcn--; } \
443 } \
444} while(0)
445
446#define MALLOC_COPY(dest,src,nbytes) \
447do { \
448 INTERNAL_SIZE_T* mcsrc = (INTERNAL_SIZE_T*) src; \
449 INTERNAL_SIZE_T* mcdst = (INTERNAL_SIZE_T*) dest; \
450 long mctmp = (nbytes)/sizeof(INTERNAL_SIZE_T), mcn; \
451 if (mctmp < 8) mcn = 0; else { mcn = (mctmp-1)/8; mctmp %= 8; } \
452 switch (mctmp) { \
453 case 0: for(;;) { *mcdst++ = *mcsrc++; \
454 case 7: *mcdst++ = *mcsrc++; \
455 case 6: *mcdst++ = *mcsrc++; \
456 case 5: *mcdst++ = *mcsrc++; \
457 case 4: *mcdst++ = *mcsrc++; \
458 case 3: *mcdst++ = *mcsrc++; \
459 case 2: *mcdst++ = *mcsrc++; \
460 case 1: *mcdst++ = *mcsrc++; if(mcn <= 0) break; mcn--; } \
461 } \
462} while(0)
463
464#endif
465
466
467/*
468 Define HAVE_MMAP to optionally make malloc() use mmap() to
469 allocate very large blocks. These will be returned to the
470 operating system immediately after a free().
471*/
472
473/***
474#ifndef HAVE_MMAP
475#define HAVE_MMAP 1
476#endif
477***/
478#undef HAVE_MMAP /* Not available for U-Boot */
479
480/*
481 Define HAVE_MREMAP to make realloc() use mremap() to re-allocate
482 large blocks. This is currently only possible on Linux with
483 kernel versions newer than 1.3.77.
484*/
485
486/***
487#ifndef HAVE_MREMAP
488#ifdef INTERNAL_LINUX_C_LIB
489#define HAVE_MREMAP 1
490#else
491#define HAVE_MREMAP 0
492#endif
493#endif
494***/
495#undef HAVE_MREMAP /* Not available for U-Boot */
496
497#if HAVE_MMAP
498
499#include <unistd.h>
500#include <fcntl.h>
501#include <sys/mman.h>
502
503#if !defined(MAP_ANONYMOUS) && defined(MAP_ANON)
504#define MAP_ANONYMOUS MAP_ANON
505#endif
506
507#endif /* HAVE_MMAP */
508
509/*
510 Access to system page size. To the extent possible, this malloc
511 manages memory from the system in page-size units.
512
513 The following mechanics for getpagesize were adapted from
514 bsd/gnu getpagesize.h
515*/
516
517#define LACKS_UNISTD_H /* Shortcut for U-Boot */
518#define malloc_getpagesize 4096
519
520#ifndef LACKS_UNISTD_H
521# include <unistd.h>
522#endif
523
524#ifndef malloc_getpagesize
525# ifdef _SC_PAGESIZE /* some SVR4 systems omit an underscore */
526# ifndef _SC_PAGE_SIZE
527# define _SC_PAGE_SIZE _SC_PAGESIZE
528# endif
529# endif
530# ifdef _SC_PAGE_SIZE
531# define malloc_getpagesize sysconf(_SC_PAGE_SIZE)
532# else
533# if defined(BSD) || defined(DGUX) || defined(HAVE_GETPAGESIZE)
534 extern size_t getpagesize();
535# define malloc_getpagesize getpagesize()
536# else
537# ifdef WIN32
538# define malloc_getpagesize (4096) /* TBD: Use 'GetSystemInfo' instead */
539# else
540# ifndef LACKS_SYS_PARAM_H
541# include <sys/param.h>
542# endif
543# ifdef EXEC_PAGESIZE
544# define malloc_getpagesize EXEC_PAGESIZE
545# else
546# ifdef NBPG
547# ifndef CLSIZE
548# define malloc_getpagesize NBPG
549# else
550# define malloc_getpagesize (NBPG * CLSIZE)
551# endif
552# else
553# ifdef NBPC
554# define malloc_getpagesize NBPC
555# else
556# ifdef PAGESIZE
557# define malloc_getpagesize PAGESIZE
558# else
559# define malloc_getpagesize (4096) /* just guess */
560# endif
561# endif
562# endif
563# endif
564# endif
565# endif
566# endif
567#endif
568
569
570
571/*
572
573 This version of malloc supports the standard SVID/XPG mallinfo
574 routine that returns a struct containing the same kind of
575 information you can get from malloc_stats. It should work on
576 any SVID/XPG compliant system that has a /usr/include/malloc.h
577 defining struct mallinfo. (If you'd like to install such a thing
578 yourself, cut out the preliminary declarations as described above
579 and below and save them in a malloc.h file. But there's no
580 compelling reason to bother to do this.)
581
582 The main declaration needed is the mallinfo struct that is returned
583 (by-copy) by mallinfo(). The SVID/XPG malloinfo struct contains a
584 bunch of fields, most of which are not even meaningful in this
585 version of malloc. Some of these fields are are instead filled by
586 mallinfo() with other numbers that might possibly be of interest.
587
588 HAVE_USR_INCLUDE_MALLOC_H should be set if you have a
589 /usr/include/malloc.h file that includes a declaration of struct
590 mallinfo. If so, it is included; else an SVID2/XPG2 compliant
591 version is declared below. These must be precisely the same for
592 mallinfo() to work.
593
594*/
595
596/* #define HAVE_USR_INCLUDE_MALLOC_H */
597
598#if HAVE_USR_INCLUDE_MALLOC_H
599#include "/usr/include/malloc.h"
600#else
601
602/* SVID2/XPG mallinfo structure */
603
604struct mallinfo {
605 int arena; /* total space allocated from system */
606 int ordblks; /* number of non-inuse chunks */
607 int smblks; /* unused -- always zero */
608 int hblks; /* number of mmapped regions */
609 int hblkhd; /* total space in mmapped regions */
610 int usmblks; /* unused -- always zero */
611 int fsmblks; /* unused -- always zero */
612 int uordblks; /* total allocated space */
613 int fordblks; /* total non-inuse space */
614 int keepcost; /* top-most, releasable (via malloc_trim) space */
615};
616
617/* SVID2/XPG mallopt options */
618
619#define M_MXFAST 1 /* UNUSED in this malloc */
620#define M_NLBLKS 2 /* UNUSED in this malloc */
621#define M_GRAIN 3 /* UNUSED in this malloc */
622#define M_KEEP 4 /* UNUSED in this malloc */
623
624#endif
625
626/* mallopt options that actually do something */
627
628#define M_TRIM_THRESHOLD -1
629#define M_TOP_PAD -2
630#define M_MMAP_THRESHOLD -3
631#define M_MMAP_MAX -4
632
633
634
635#ifndef DEFAULT_TRIM_THRESHOLD
636#define DEFAULT_TRIM_THRESHOLD (128 * 1024)
637#endif
638
639/*
640 M_TRIM_THRESHOLD is the maximum amount of unused top-most memory
641 to keep before releasing via malloc_trim in free().
642
643 Automatic trimming is mainly useful in long-lived programs.
644 Because trimming via sbrk can be slow on some systems, and can
645 sometimes be wasteful (in cases where programs immediately
646 afterward allocate more large chunks) the value should be high
647 enough so that your overall system performance would improve by
648 releasing.
649
650 The trim threshold and the mmap control parameters (see below)
651 can be traded off with one another. Trimming and mmapping are
652 two different ways of releasing unused memory back to the
653 system. Between these two, it is often possible to keep
654 system-level demands of a long-lived program down to a bare
655 minimum. For example, in one test suite of sessions measuring
656 the XF86 X server on Linux, using a trim threshold of 128K and a
657 mmap threshold of 192K led to near-minimal long term resource
658 consumption.
659
660 If you are using this malloc in a long-lived program, it should
661 pay to experiment with these values. As a rough guide, you
662 might set to a value close to the average size of a process
663 (program) running on your system. Releasing this much memory
664 would allow such a process to run in memory. Generally, it's
665 worth it to tune for trimming rather tham memory mapping when a
666 program undergoes phases where several large chunks are
667 allocated and released in ways that can reuse each other's
668 storage, perhaps mixed with phases where there are no such
669 chunks at all. And in well-behaved long-lived programs,
670 controlling release of large blocks via trimming versus mapping
671 is usually faster.
672
673 However, in most programs, these parameters serve mainly as
674 protection against the system-level effects of carrying around
675 massive amounts of unneeded memory. Since frequent calls to
676 sbrk, mmap, and munmap otherwise degrade performance, the default
677 parameters are set to relatively high values that serve only as
678 safeguards.
679
680 The default trim value is high enough to cause trimming only in
681 fairly extreme (by current memory consumption standards) cases.
682 It must be greater than page size to have any useful effect. To
683 disable trimming completely, you can set to (unsigned long)(-1);
684
685
686*/
687
688
689#ifndef DEFAULT_TOP_PAD
690#define DEFAULT_TOP_PAD (0)
691#endif
692
693/*
694 M_TOP_PAD is the amount of extra `padding' space to allocate or
695 retain whenever sbrk is called. It is used in two ways internally:
696
697 * When sbrk is called to extend the top of the arena to satisfy
698 a new malloc request, this much padding is added to the sbrk
699 request.
700
701 * When malloc_trim is called automatically from free(),
702 it is used as the `pad' argument.
703
704 In both cases, the actual amount of padding is rounded
705 so that the end of the arena is always a system page boundary.
706
707 The main reason for using padding is to avoid calling sbrk so
708 often. Having even a small pad greatly reduces the likelihood
709 that nearly every malloc request during program start-up (or
710 after trimming) will invoke sbrk, which needlessly wastes
711 time.
712
713 Automatic rounding-up to page-size units is normally sufficient
714 to avoid measurable overhead, so the default is 0. However, in
715 systems where sbrk is relatively slow, it can pay to increase
716 this value, at the expense of carrying around more memory than
717 the program needs.
718
719*/
720
721
722#ifndef DEFAULT_MMAP_THRESHOLD
723#define DEFAULT_MMAP_THRESHOLD (128 * 1024)
724#endif
725
726/*
727
728 M_MMAP_THRESHOLD is the request size threshold for using mmap()
729 to service a request. Requests of at least this size that cannot
730 be allocated using already-existing space will be serviced via mmap.
731 (If enough normal freed space already exists it is used instead.)
732
733 Using mmap segregates relatively large chunks of memory so that
734 they can be individually obtained and released from the host
735 system. A request serviced through mmap is never reused by any
736 other request (at least not directly; the system may just so
737 happen to remap successive requests to the same locations).
738
739 Segregating space in this way has the benefit that mmapped space
740 can ALWAYS be individually released back to the system, which
741 helps keep the system level memory demands of a long-lived
742 program low. Mapped memory can never become `locked' between
743 other chunks, as can happen with normally allocated chunks, which
744 menas that even trimming via malloc_trim would not release them.
745
746 However, it has the disadvantages that:
747
748 1. The space cannot be reclaimed, consolidated, and then
749 used to service later requests, as happens with normal chunks.
750 2. It can lead to more wastage because of mmap page alignment
751 requirements
752 3. It causes malloc performance to be more dependent on host
753 system memory management support routines which may vary in
754 implementation quality and may impose arbitrary
755 limitations. Generally, servicing a request via normal
756 malloc steps is faster than going through a system's mmap.
757
758 All together, these considerations should lead you to use mmap
759 only for relatively large requests.
760
761
762*/
763
764
765
766#ifndef DEFAULT_MMAP_MAX
767#if HAVE_MMAP
768#define DEFAULT_MMAP_MAX (64)
769#else
770#define DEFAULT_MMAP_MAX (0)
771#endif
772#endif
773
774/*
775 M_MMAP_MAX is the maximum number of requests to simultaneously
776 service using mmap. This parameter exists because:
777
778 1. Some systems have a limited number of internal tables for
779 use by mmap.
780 2. In most systems, overreliance on mmap can degrade overall
781 performance.
782 3. If a program allocates many large regions, it is probably
783 better off using normal sbrk-based allocation routines that
784 can reclaim and reallocate normal heap memory. Using a
785 small value allows transition into this mode after the
786 first few allocations.
787
788 Setting to 0 disables all use of mmap. If HAVE_MMAP is not set,
789 the default value is 0, and attempts to set it to non-zero values
790 in mallopt will fail.
791*/
792
793
794
795
796/*
797 USE_DL_PREFIX will prefix all public routines with the string 'dl'.
798 Useful to quickly avoid procedure declaration conflicts and linker
799 symbol conflicts with existing memory allocation routines.
800
801*/
802
803/* #define USE_DL_PREFIX */
804
805
806
807
808/*
809
810 Special defines for linux libc
811
812 Except when compiled using these special defines for Linux libc
813 using weak aliases, this malloc is NOT designed to work in
814 multithreaded applications. No semaphores or other concurrency
815 control are provided to ensure that multiple malloc or free calls
816 don't run at the same time, which could be disasterous. A single
817 semaphore could be used across malloc, realloc, and free (which is
818 essentially the effect of the linux weak alias approach). It would
819 be hard to obtain finer granularity.
820
821*/
822
823
824#ifdef INTERNAL_LINUX_C_LIB
825
826#if __STD_C
827
828Void_t * __default_morecore_init (ptrdiff_t);
829Void_t *(*__morecore)(ptrdiff_t) = __default_morecore_init;
830
831#else
832
833Void_t * __default_morecore_init ();
834Void_t *(*__morecore)() = __default_morecore_init;
835
836#endif
837
838#define MORECORE (*__morecore)
839#define MORECORE_FAILURE 0
840#define MORECORE_CLEARS 1
841
842#else /* INTERNAL_LINUX_C_LIB */
843
844#if __STD_C
845extern Void_t* sbrk(ptrdiff_t);
846#else
847extern Void_t* sbrk();
848#endif
849
850#ifndef MORECORE
851#define MORECORE sbrk
852#endif
853
854#ifndef MORECORE_FAILURE
855#define MORECORE_FAILURE -1
856#endif
857
858#ifndef MORECORE_CLEARS
859#define MORECORE_CLEARS 1
860#endif
861
862#endif /* INTERNAL_LINUX_C_LIB */
863
864#if defined(INTERNAL_LINUX_C_LIB) && defined(__ELF__)
865
866#define cALLOc __libc_calloc
867#define fREe __libc_free
868#define mALLOc __libc_malloc
869#define mEMALIGn __libc_memalign
870#define rEALLOc __libc_realloc
871#define vALLOc __libc_valloc
872#define pvALLOc __libc_pvalloc
873#define mALLINFo __libc_mallinfo
874#define mALLOPt __libc_mallopt
875
876#pragma weak calloc = __libc_calloc
877#pragma weak free = __libc_free
878#pragma weak cfree = __libc_free
879#pragma weak malloc = __libc_malloc
880#pragma weak memalign = __libc_memalign
881#pragma weak realloc = __libc_realloc
882#pragma weak valloc = __libc_valloc
883#pragma weak pvalloc = __libc_pvalloc
884#pragma weak mallinfo = __libc_mallinfo
885#pragma weak mallopt = __libc_mallopt
886
887#else
888
889#ifdef USE_DL_PREFIX
890#define cALLOc dlcalloc
891#define fREe dlfree
892#define mALLOc dlmalloc
893#define mEMALIGn dlmemalign
894#define rEALLOc dlrealloc
895#define vALLOc dlvalloc
896#define pvALLOc dlpvalloc
897#define mALLINFo dlmallinfo
898#define mALLOPt dlmallopt
899#else /* USE_DL_PREFIX */
900#define cALLOc calloc
901#define fREe free
902#define mALLOc malloc
903#define mEMALIGn memalign
904#define rEALLOc realloc
905#define vALLOc valloc
906#define pvALLOc pvalloc
907#define mALLINFo mallinfo
908#define mALLOPt mallopt
909#endif /* USE_DL_PREFIX */
910
911#endif
912
913/* Public routines */
914
915#if __STD_C
916
917Void_t* mALLOc(size_t);
918void fREe(Void_t*);
919Void_t* rEALLOc(Void_t*, size_t);
920Void_t* mEMALIGn(size_t, size_t);
921Void_t* vALLOc(size_t);
922Void_t* pvALLOc(size_t);
923Void_t* cALLOc(size_t, size_t);
924void cfree(Void_t*);
925int malloc_trim(size_t);
926size_t malloc_usable_size(Void_t*);
927void malloc_stats(void);
928int mALLOPt(int, int);
929struct mallinfo mALLINFo(void);
930#else
931Void_t* mALLOc();
932void fREe();
933Void_t* rEALLOc();
934Void_t* mEMALIGn();
935Void_t* vALLOc();
936Void_t* pvALLOc();
937Void_t* cALLOc();
938void cfree();
939int malloc_trim();
940size_t malloc_usable_size();
941void malloc_stats();
942int mALLOPt();
943struct mallinfo mALLINFo();
944#endif
945
946
947#ifdef __cplusplus
948}; /* end of extern "C" */
949#endif