Home | History | Annotate | Line # | Download | only in internal
      1 #ifndef JEMALLOC_INTERNAL_BIN_H
      2 #define JEMALLOC_INTERNAL_BIN_H
      3 
      4 #include "jemalloc/internal/extent_types.h"
      5 #include "jemalloc/internal/extent_structs.h"
      6 #include "jemalloc/internal/mutex.h"
      7 #include "jemalloc/internal/bin_stats.h"
      8 
      9 /*
     10  * A bin contains a set of extents that are currently being used for slab
     11  * allocations.
     12  */
     13 
     14 /*
     15  * Read-only information associated with each element of arena_t's bins array
     16  * is stored separately, partly to reduce memory usage (only one copy, rather
     17  * than one per arena), but mainly to avoid false cacheline sharing.
     18  *
     19  * Each slab has the following layout:
     20  *
     21  *   /--------------------\
     22  *   | region 0           |
     23  *   |--------------------|
     24  *   | region 1           |
     25  *   |--------------------|
     26  *   | ...                |
     27  *   | ...                |
     28  *   | ...                |
     29  *   |--------------------|
     30  *   | region nregs-1     |
     31  *   \--------------------/
     32  */
     33 typedef struct bin_info_s bin_info_t;
     34 struct bin_info_s {
     35 	/* Size of regions in a slab for this bin's size class. */
     36 	size_t			reg_size;
     37 
     38 	/* Total size of a slab for this bin's size class. */
     39 	size_t			slab_size;
     40 
     41 	/* Total number of regions in a slab for this bin's size class. */
     42 	uint32_t		nregs;
     43 
     44 	/*
     45 	 * Metadata used to manipulate bitmaps for slabs associated with this
     46 	 * bin.
     47 	 */
     48 	bitmap_info_t		bitmap_info;
     49 };
     50 
     51 extern const bin_info_t bin_infos[NBINS];
     52 
     53 
     54 typedef struct bin_s bin_t;
     55 struct bin_s {
     56 	/* All operations on bin_t fields require lock ownership. */
     57 	malloc_mutex_t		lock;
     58 
     59 	/*
     60 	 * Current slab being used to service allocations of this bin's size
     61 	 * class.  slabcur is independent of slabs_{nonfull,full}; whenever
     62 	 * slabcur is reassigned, the previous slab must be deallocated or
     63 	 * inserted into slabs_{nonfull,full}.
     64 	 */
     65 	extent_t		*slabcur;
     66 
     67 	/*
     68 	 * Heap of non-full slabs.  This heap is used to assure that new
     69 	 * allocations come from the non-full slab that is oldest/lowest in
     70 	 * memory.
     71 	 */
     72 	extent_heap_t		slabs_nonfull;
     73 
     74 	/* List used to track full slabs. */
     75 	extent_list_t		slabs_full;
     76 
     77 	/* Bin statistics. */
     78 	bin_stats_t	stats;
     79 };
     80 
     81 /* Initializes a bin to empty.  Returns true on error. */
     82 bool bin_init(bin_t *bin);
     83 
     84 /* Forking. */
     85 void bin_prefork(tsdn_t *tsdn, bin_t *bin);
     86 void bin_postfork_parent(tsdn_t *tsdn, bin_t *bin);
     87 void bin_postfork_child(tsdn_t *tsdn, bin_t *bin);
     88 
     89 /* Stats. */
     90 static inline void
     91 bin_stats_merge(tsdn_t *tsdn, bin_stats_t *dst_bin_stats, bin_t *bin) {
     92 	malloc_mutex_lock(tsdn, &bin->lock);
     93 	malloc_mutex_prof_read(tsdn, &dst_bin_stats->mutex_data, &bin->lock);
     94 	dst_bin_stats->nmalloc += bin->stats.nmalloc;
     95 	dst_bin_stats->ndalloc += bin->stats.ndalloc;
     96 	dst_bin_stats->nrequests += bin->stats.nrequests;
     97 	dst_bin_stats->curregs += bin->stats.curregs;
     98 	dst_bin_stats->nfills += bin->stats.nfills;
     99 	dst_bin_stats->nflushes += bin->stats.nflushes;
    100 	dst_bin_stats->nslabs += bin->stats.nslabs;
    101 	dst_bin_stats->reslabs += bin->stats.reslabs;
    102 	dst_bin_stats->curslabs += bin->stats.curslabs;
    103 	malloc_mutex_unlock(tsdn, &bin->lock);
    104 }
    105 
    106 #endif /* JEMALLOC_INTERNAL_BIN_H */
    107