tor

The Tor anonymity network
git clone https://git.dasho.dev/tor.git
Log | Files | Refs | README | LICENSE

memarea.c (11535B)


      1 /* Copyright (c) 2008-2021, The Tor Project, Inc. */
      2 /* See LICENSE for licensing information */
      3 
      4 /**
      5 * \file memarea.c
      6 *
      7 * \brief Implementation for memarea_t, an allocator for allocating lots of
      8 * small objects that will be freed all at once.
      9 */
     10 
     11 #include "orconfig.h"
     12 #include "lib/memarea/memarea.h"
     13 
     14 #include <stdlib.h>
     15 #include <string.h>
     16 
     17 #include "lib/arch/bytes.h"
     18 #include "lib/cc/torint.h"
     19 #include "lib/smartlist_core/smartlist_core.h"
     20 #include "lib/smartlist_core/smartlist_foreach.h"
     21 #include "lib/log/log.h"
     22 #include "lib/log/util_bug.h"
     23 #include "lib/malloc/malloc.h"
     24 
     25 #ifndef DISABLE_MEMORY_SENTINELS
     26 
     27 /** If true, we try to detect any attempts to write beyond the length of a
     28 * memarea. */
     29 #define USE_SENTINELS
     30 
     31 /** All returned pointers should be aligned to the nearest multiple of this
     32 * value. */
     33 #define MEMAREA_ALIGN SIZEOF_VOID_P
     34 
     35 /** A value which, when masked out of a pointer, produces a maximally aligned
     36 * pointer. */
     37 #if MEMAREA_ALIGN == 4
     38 #define MEMAREA_ALIGN_MASK ((uintptr_t)3)
     39 #elif MEMAREA_ALIGN == 8
     40 #define MEMAREA_ALIGN_MASK ((uintptr_t)7)
     41 #else
     42 #error "void* is neither 4 nor 8 bytes long."
     43 #endif /* MEMAREA_ALIGN == 4 || ... */
     44 
     45 #if defined(__GNUC__) && defined(FLEXIBLE_ARRAY_MEMBER)
     46 #define USE_ALIGNED_ATTRIBUTE
     47 /** Name for the 'memory' member of a memory chunk. */
     48 #define U_MEM mem
     49 #else
     50 #define U_MEM u.mem
     51 #endif /* defined(__GNUC__) && defined(FLEXIBLE_ARRAY_MEMBER) */
     52 
     53 #ifdef USE_SENTINELS
     54 /** Magic value that we stick at the end of a memarea so we can make sure
     55 * there are no run-off-the-end bugs. */
     56 #define SENTINEL_VAL 0x90806622u
     57 /** How many bytes per area do we devote to the sentinel? */
     58 #define SENTINEL_LEN sizeof(uint32_t)
     59 /** Given a mem_area_chunk_t with SENTINEL_LEN extra bytes allocated at the
     60 * end, set those bytes. */
     61 #define SET_SENTINEL(chunk)                                     \
     62  STMT_BEGIN                                                    \
     63  set_uint32( &(chunk)->U_MEM[chunk->mem_size], SENTINEL_VAL ); \
     64  STMT_END
     65 /** Assert that the sentinel on a memarea is set correctly. */
     66 #define CHECK_SENTINEL(chunk)                                           \
     67  STMT_BEGIN                                                            \
     68  uint32_t sent_val = get_uint32(&(chunk)->U_MEM[chunk->mem_size]);     \
     69  tor_assert(sent_val == SENTINEL_VAL);                                 \
     70  STMT_END
     71 #else /* !defined(USE_SENTINELS) */
     72 #define SENTINEL_LEN 0
     73 #define SET_SENTINEL(chunk) STMT_NIL
     74 #define CHECK_SENTINEL(chunk) STMT_NIL
     75 #endif /* defined(USE_SENTINELS) */
     76 
     77 /** Increment <b>ptr</b> until it is aligned to MEMAREA_ALIGN. */
     78 static inline void *
     79 realign_pointer(void *ptr)
     80 {
     81  uintptr_t x = (uintptr_t)ptr;
     82  x = (x+MEMAREA_ALIGN_MASK) & ~MEMAREA_ALIGN_MASK;
     83  /* Reinstate this if bug 930 ever reappears
     84  tor_assert(((void*)x) >= ptr);
     85  */
     86  return (void*)x;
     87 }
     88 
     89 /** Implements part of a memarea.  New memory is carved off from chunk->mem in
     90 * increasing order until a request is too big, at which point a new chunk is
     91 * allocated. */
     92 typedef struct memarea_chunk_t {
     93  /** Next chunk in this area. Only kept around so we can free it. */
     94  struct memarea_chunk_t *next_chunk;
     95  size_t mem_size; /**< How much RAM is available in mem, total? */
     96  char *next_mem; /**< Next position in mem to allocate data at.  If it's
     97                   * equal to mem+mem_size, this chunk is full. */
     98 #ifdef USE_ALIGNED_ATTRIBUTE
     99  /** Actual content of the memory chunk. */
    100  char mem[FLEXIBLE_ARRAY_MEMBER] __attribute__((aligned(MEMAREA_ALIGN)));
    101 #else
    102  union {
    103    char mem[1]; /**< Memory space in this chunk.  */
    104    void *void_for_alignment_; /**< Dummy; used to make sure mem is aligned. */
    105  } u; /**< Union used to enforce alignment when we don't have support for
    106        * doing it right. */
    107 #endif /* defined(USE_ALIGNED_ATTRIBUTE) */
    108 } memarea_chunk_t;
    109 
    110 /** How many bytes are needed for overhead before we get to the memory part
    111 * of a chunk? */
    112 #define CHUNK_HEADER_SIZE offsetof(memarea_chunk_t, U_MEM)
    113 
    114 /** What's the smallest that we'll allocate a chunk? */
    115 #define CHUNK_SIZE 4096
    116 
    117 /** A memarea_t is an allocation region for a set of small memory requests
    118 * that will all be freed at once. */
    119 struct memarea_t {
    120  memarea_chunk_t *first; /**< Top of the chunk stack: never NULL. */
    121 };
    122 
    123 /** Helper: allocate a new memarea chunk of around <b>chunk_size</b> bytes. */
    124 static memarea_chunk_t *
    125 alloc_chunk(size_t sz)
    126 {
    127  tor_assert(sz < SIZE_T_CEILING);
    128 
    129  size_t chunk_size = sz < CHUNK_SIZE ? CHUNK_SIZE : sz;
    130  memarea_chunk_t *res;
    131  chunk_size += SENTINEL_LEN;
    132  res = tor_malloc(chunk_size);
    133  res->next_chunk = NULL;
    134  res->mem_size = chunk_size - CHUNK_HEADER_SIZE - SENTINEL_LEN;
    135  res->next_mem = res->U_MEM;
    136  tor_assert(res->next_mem+res->mem_size+SENTINEL_LEN ==
    137             ((char*)res)+chunk_size);
    138  tor_assert(realign_pointer(res->next_mem) == res->next_mem);
    139  SET_SENTINEL(res);
    140  return res;
    141 }
    142 
    143 /** Release <b>chunk</b> from a memarea. */
    144 static void
    145 memarea_chunk_free_unchecked(memarea_chunk_t *chunk)
    146 {
    147  CHECK_SENTINEL(chunk);
    148  tor_free(chunk);
    149 }
    150 
    151 /** Allocate and return new memarea. */
    152 memarea_t *
    153 memarea_new(void)
    154 {
    155  memarea_t *head = tor_malloc(sizeof(memarea_t));
    156  head->first = alloc_chunk(CHUNK_SIZE);
    157  return head;
    158 }
    159 
    160 /** Free <b>area</b>, invalidating all pointers returned from memarea_alloc()
    161 * and friends for this area */
    162 void
    163 memarea_drop_all_(memarea_t *area)
    164 {
    165  memarea_chunk_t *chunk, *next;
    166  for (chunk = area->first; chunk; chunk = next) {
    167    next = chunk->next_chunk;
    168    memarea_chunk_free_unchecked(chunk);
    169  }
    170  area->first = NULL; /*fail fast on */
    171  tor_free(area);
    172 }
    173 
    174 /** Forget about having allocated anything in <b>area</b>, and free some of
    175 * the backing storage associated with it, as appropriate. Invalidates all
    176 * pointers returned from memarea_alloc() for this area. */
    177 void
    178 memarea_clear(memarea_t *area)
    179 {
    180  memarea_chunk_t *chunk, *next;
    181  if (area->first->next_chunk) {
    182    for (chunk = area->first->next_chunk; chunk; chunk = next) {
    183      next = chunk->next_chunk;
    184      memarea_chunk_free_unchecked(chunk);
    185    }
    186    area->first->next_chunk = NULL;
    187  }
    188  area->first->next_mem = area->first->U_MEM;
    189 }
    190 
    191 /** Return true iff <b>p</b> is in a range that has been returned by an
    192 * allocation from <b>area</b>. */
    193 int
    194 memarea_owns_ptr(const memarea_t *area, const void *p)
    195 {
    196  memarea_chunk_t *chunk;
    197  const char *ptr = p;
    198  for (chunk = area->first; chunk; chunk = chunk->next_chunk) {
    199    if (ptr >= chunk->U_MEM && ptr < chunk->next_mem)
    200      return 1;
    201  }
    202  return 0;
    203 }
    204 
    205 /** Return a pointer to a chunk of memory in <b>area</b> of at least <b>sz</b>
    206 * bytes.  <b>sz</b> should be significantly smaller than the area's chunk
    207 * size, though we can deal if it isn't. */
    208 void *
    209 memarea_alloc(memarea_t *area, size_t sz)
    210 {
    211  memarea_chunk_t *chunk = area->first;
    212  char *result;
    213  tor_assert(chunk);
    214  CHECK_SENTINEL(chunk);
    215  tor_assert(sz < SIZE_T_CEILING);
    216  if (sz == 0)
    217    sz = 1;
    218  tor_assert(chunk->next_mem <= chunk->U_MEM + chunk->mem_size);
    219  const size_t space_remaining =
    220    (chunk->U_MEM + chunk->mem_size) - chunk->next_mem;
    221  if (sz > space_remaining) {
    222    if (sz+CHUNK_HEADER_SIZE >= CHUNK_SIZE) {
    223      /* This allocation is too big.  Stick it in a special chunk, and put
    224       * that chunk second in the list. */
    225      memarea_chunk_t *new_chunk = alloc_chunk(sz+CHUNK_HEADER_SIZE);
    226      new_chunk->next_chunk = chunk->next_chunk;
    227      chunk->next_chunk = new_chunk;
    228      chunk = new_chunk;
    229    } else {
    230      memarea_chunk_t *new_chunk = alloc_chunk(CHUNK_SIZE);
    231      new_chunk->next_chunk = chunk;
    232      area->first = chunk = new_chunk;
    233    }
    234    tor_assert(chunk->mem_size >= sz);
    235  }
    236  result = chunk->next_mem;
    237  chunk->next_mem = chunk->next_mem + sz;
    238  /* Reinstate these if bug 930 ever comes back
    239  tor_assert(chunk->next_mem >= chunk->U_MEM);
    240  tor_assert(chunk->next_mem <= chunk->U_MEM+chunk->mem_size);
    241  */
    242  chunk->next_mem = realign_pointer(chunk->next_mem);
    243  return result;
    244 }
    245 
    246 /** As memarea_alloc(), but clears the memory it returns. */
    247 void *
    248 memarea_alloc_zero(memarea_t *area, size_t sz)
    249 {
    250  void *result = memarea_alloc(area, sz);
    251  memset(result, 0, sz);
    252  return result;
    253 }
    254 
    255 /** As memdup, but returns the memory from <b>area</b>. */
    256 void *
    257 memarea_memdup(memarea_t *area, const void *s, size_t n)
    258 {
    259  char *result = memarea_alloc(area, n);
    260  memcpy(result, s, n);
    261  return result;
    262 }
    263 
    264 /** As strdup, but returns the memory from <b>area</b>. */
    265 char *
    266 memarea_strdup(memarea_t *area, const char *s)
    267 {
    268  return memarea_memdup(area, s, strlen(s)+1);
    269 }
    270 
    271 /** As strndup, but returns the memory from <b>area</b>. */
    272 char *
    273 memarea_strndup(memarea_t *area, const char *s, size_t n)
    274 {
    275  size_t ln = 0;
    276  char *result;
    277  tor_assert(n < SIZE_T_CEILING);
    278  for (ln = 0; ln < n && s[ln]; ++ln)
    279    ;
    280  result = memarea_alloc(area, ln+1);
    281  memcpy(result, s, ln);
    282  result[ln]='\0';
    283  return result;
    284 }
    285 
    286 /** Set <b>allocated_out</b> to the number of bytes allocated in <b>area</b>,
    287 * and <b>used_out</b> to the number of bytes currently used. */
    288 void
    289 memarea_get_stats(memarea_t *area, size_t *allocated_out, size_t *used_out)
    290 {
    291  size_t a = 0, u = 0;
    292  memarea_chunk_t *chunk;
    293  for (chunk = area->first; chunk; chunk = chunk->next_chunk) {
    294    CHECK_SENTINEL(chunk);
    295    a += CHUNK_HEADER_SIZE + chunk->mem_size;
    296    tor_assert(chunk->next_mem >= chunk->U_MEM);
    297    u += CHUNK_HEADER_SIZE + (chunk->next_mem - chunk->U_MEM);
    298  }
    299  *allocated_out = a;
    300  *used_out = u;
    301 }
    302 
    303 /** Assert that <b>area</b> is okay. */
    304 void
    305 memarea_assert_ok(memarea_t *area)
    306 {
    307  memarea_chunk_t *chunk;
    308  tor_assert(area->first);
    309 
    310  for (chunk = area->first; chunk; chunk = chunk->next_chunk) {
    311    CHECK_SENTINEL(chunk);
    312    tor_assert(chunk->next_mem >= chunk->U_MEM);
    313    tor_assert(chunk->next_mem <=
    314          (char*) realign_pointer(chunk->U_MEM+chunk->mem_size));
    315  }
    316 }
    317 
    318 #else /* defined(DISABLE_MEMORY_SENTINELS) */
    319 
    320 struct memarea_t {
    321  smartlist_t *pieces;
    322 };
    323 
    324 memarea_t *
    325 memarea_new(void)
    326 {
    327  memarea_t *ma = tor_malloc_zero(sizeof(memarea_t));
    328  ma->pieces = smartlist_new();
    329  return ma;
    330 }
    331 void
    332 memarea_drop_all_(memarea_t *area)
    333 {
    334  memarea_clear(area);
    335  smartlist_free(area->pieces);
    336  tor_free(area);
    337 }
    338 void
    339 memarea_clear(memarea_t *area)
    340 {
    341  SMARTLIST_FOREACH(area->pieces, void *, p, tor_free_(p));
    342  smartlist_clear(area->pieces);
    343 }
    344 int
    345 memarea_owns_ptr(const memarea_t *area, const void *ptr)
    346 {
    347  SMARTLIST_FOREACH(area->pieces, const void *, p, if (ptr == p) return 1;);
    348  return 0;
    349 }
    350 
    351 void *
    352 memarea_alloc(memarea_t *area, size_t sz)
    353 {
    354  void *result = tor_malloc(sz);
    355  smartlist_add(area->pieces, result);
    356  return result;
    357 }
    358 
    359 void *
    360 memarea_alloc_zero(memarea_t *area, size_t sz)
    361 {
    362  void *result = tor_malloc_zero(sz);
    363  smartlist_add(area->pieces, result);
    364  return result;
    365 }
    366 void *
    367 memarea_memdup(memarea_t *area, const void *s, size_t n)
    368 {
    369  void *r = memarea_alloc(area, n);
    370  memcpy(r, s, n);
    371  return r;
    372 }
    373 char *
    374 memarea_strdup(memarea_t *area, const char *s)
    375 {
    376  size_t n = strlen(s);
    377  char *r = memarea_alloc(area, n+1);
    378  memcpy(r, s, n);
    379  r[n] = 0;
    380  return r;
    381 }
    382 char *
    383 memarea_strndup(memarea_t *area, const char *s, size_t n)
    384 {
    385  size_t ln = strnlen(s, n);
    386  char *r = memarea_alloc(area, ln+1);
    387  memcpy(r, s, ln);
    388  r[ln] = 0;
    389  return r;
    390 }
    391 void
    392 memarea_get_stats(memarea_t *area,
    393                  size_t *allocated_out, size_t *used_out)
    394 {
    395  (void)area;
    396  *allocated_out = *used_out = 128;
    397 }
    398 void
    399 memarea_assert_ok(memarea_t *area)
    400 {
    401  (void)area;
    402 }
    403 
    404 #endif /* !defined(DISABLE_MEMORY_SENTINELS) */