X-Git-Url: https://gerrit.fd.io/r/gitweb?a=blobdiff_plain;f=src%2Fvppinfra%2Fmem.h;h=75015d59a4ad3fc9383c8117833e70df391e4093;hb=dd298e804a25dbebc9e284b3b2d0dfd9bf674ad8;hp=2fd4bfb5dbb44579cd6b91429020579d605c17c2;hpb=4537c30925050ffa34c33e6a481f07f1ec0a01ff;p=vpp.git diff --git a/src/vppinfra/mem.h b/src/vppinfra/mem.h index 2fd4bfb5dbb..75015d59a4a 100644 --- a/src/vppinfra/mem.h +++ b/src/vppinfra/mem.h @@ -47,12 +47,16 @@ #include #include /* memcpy, clib_memset */ -#include +#ifdef CLIB_SANITIZE_ADDR +#include +#endif #define CLIB_MAX_MHEAPS 256 #define CLIB_MAX_NUMAS 16 #define CLIB_MEM_VM_MAP_FAILED ((void *) ~0) #define CLIB_MEM_ERROR (-1) +#define CLIB_MEM_LOG2_MIN_ALIGN (3) +#define CLIB_MEM_MIN_ALIGN (1 << CLIB_MEM_LOG2_MIN_ALIGN) typedef enum { @@ -93,14 +97,50 @@ typedef struct _clib_mem_vm_map_hdr struct _clib_mem_vm_map_hdr *prev, *next; } clib_mem_vm_map_hdr_t; +#define foreach_clib_mem_heap_flag \ + _ (0, LOCKED, "locked") \ + _ (1, UNMAP_ON_DESTROY, "unmap-on-destroy") \ + _ (2, TRACED, "traced") + +typedef enum +{ +#define _(i, v, s) CLIB_MEM_HEAP_F_##v = (1 << i), + foreach_clib_mem_heap_flag +#undef _ +} clib_mem_heap_flag_t; + +typedef struct +{ + /* base address */ + void *base; + + /* dlmalloc mspace */ + void *mspace; + + /* heap size */ + uword size; + + /* page size (log2) */ + clib_mem_page_sz_t log2_page_sz:8; + + /* flags */ + clib_mem_heap_flag_t flags:8; + + /* name - _MUST_ be last */ + char name[0]; +} clib_mem_heap_t; + typedef struct { /* log2 system page size */ clib_mem_page_sz_t log2_page_sz; - /* log2 system default hugepage size */ + /* log2 default hugepage size */ clib_mem_page_sz_t log2_default_hugepage_sz; + /* log2 system default hugepage size */ + clib_mem_page_sz_t log2_sys_default_hugepage_sz; + /* bitmap of available numa nodes */ u32 numa_node_bitmap; @@ -113,6 +153,9 @@ typedef struct /* memory maps */ clib_mem_vm_map_hdr_t *first_map, *last_map; + /* map lock */ + u8 map_lock; + /* last error */ clib_error_t *error; } clib_mem_main_t; @@ -122,7 +165,23 @@ extern clib_mem_main_t clib_mem_main; /* Unspecified NUMA socket */ #define VEC_NUMA_UNSPECIFIED (0xFF) -always_inline void * +static_always_inline void +clib_mem_poison (const void volatile *p, uword s) +{ +#ifdef CLIB_SANITIZE_ADDR + ASAN_POISON_MEMORY_REGION (p, s); +#endif +} + +static_always_inline void +clib_mem_unpoison (const void volatile *p, uword s) +{ +#ifdef CLIB_SANITIZE_ADDR + ASAN_UNPOISON_MEMORY_REGION (p, s); +#endif +} + +always_inline clib_mem_heap_t * clib_mem_get_per_cpu_heap (void) { int cpu = os_get_thread_index (); @@ -130,7 +189,7 @@ clib_mem_get_per_cpu_heap (void) } always_inline void * -clib_mem_set_per_cpu_heap (u8 * new_heap) +clib_mem_set_per_cpu_heap (void *new_heap) { int cpu = os_get_thread_index (); void *old = clib_mem_main.per_cpu_mheaps[cpu]; @@ -146,7 +205,7 @@ clib_mem_get_per_numa_heap (u32 numa_id) } always_inline void * -clib_mem_set_per_numa_heap (u8 * new_heap) +clib_mem_set_per_numa_heap (void *new_heap) { int numa = os_get_numa_index (); void *old = clib_mem_main.per_numa_mheaps[numa]; @@ -175,80 +234,29 @@ clib_mem_set_thread_index (void) ASSERT (__os_thread_index > 0); } -always_inline uword -clib_mem_size_nocheck (void *p) -{ - size_t mspace_usable_size_with_delta (const void *p); - return mspace_usable_size_with_delta (p); -} - -/* Memory allocator which may call os_out_of_memory() if it fails */ -always_inline void * -clib_mem_alloc_aligned_at_offset (uword size, uword align, uword align_offset, - int os_out_of_memory_on_failure) -{ - void *heap, *p; - uword cpu; - void *mspace_get_aligned (void *msp, unsigned long n_user_data_bytes, - unsigned long align, unsigned long align_offset); - - if (align_offset > align) - { - if (align > 0) - align_offset %= align; - else - align_offset = align; - } - - cpu = os_get_thread_index (); - heap = clib_mem_main.per_cpu_mheaps[cpu]; - - p = mspace_get_aligned (heap, size, align, align_offset); - - if (PREDICT_FALSE (0 == p)) - { - if (os_out_of_memory_on_failure) - os_out_of_memory (); - return 0; - } - - CLIB_MEM_UNPOISON (p, size); - return p; -} - /* Memory allocator which calls os_out_of_memory() when it fails */ -always_inline void * -clib_mem_alloc (uword size) -{ - return clib_mem_alloc_aligned_at_offset (size, /* align */ 1, - /* align_offset */ 0, - /* os_out_of_memory */ 1); -} - -always_inline void * -clib_mem_alloc_aligned (uword size, uword align) -{ - return clib_mem_alloc_aligned_at_offset (size, align, /* align_offset */ 0, - /* os_out_of_memory */ 1); -} - -/* Memory allocator which calls os_out_of_memory() when it fails */ -always_inline void * -clib_mem_alloc_or_null (uword size) -{ - return clib_mem_alloc_aligned_at_offset (size, /* align */ 1, - /* align_offset */ 0, - /* os_out_of_memory */ 0); -} - -always_inline void * -clib_mem_alloc_aligned_or_null (uword size, uword align) -{ - return clib_mem_alloc_aligned_at_offset (size, align, /* align_offset */ 0, - /* os_out_of_memory */ 0); -} - - +void *clib_mem_alloc (uword size); +void *clib_mem_alloc_aligned (uword size, uword align); +void *clib_mem_alloc_or_null (uword size); +void *clib_mem_alloc_aligned_or_null (uword size, uword align); +void *clib_mem_realloc (void *p, uword new_size); +void *clib_mem_realloc_aligned (void *p, uword new_size, uword align); +uword clib_mem_is_heap_object (void *p); +void clib_mem_free (void *p); + +void *clib_mem_heap_alloc (void *heap, uword size); +void *clib_mem_heap_alloc_aligned (void *heap, uword size, uword align); +void *clib_mem_heap_alloc_or_null (void *heap, uword size); +void *clib_mem_heap_alloc_aligned_or_null (void *heap, uword size, + uword align); +void *clib_mem_heap_realloc (void *heap, void *p, uword new_size); +void *clib_mem_heap_realloc_aligned (void *heap, void *p, uword new_size, + uword align); +uword clib_mem_heap_is_heap_object (void *heap, void *p); +void clib_mem_heap_free (void *heap, void *p); + +uword clib_mem_size (void *p); +void clib_mem_free_s (void *p); /* Memory allocator which panics when it fails. Use macro so that clib_panic macro can expand __FUNCTION__ and __LINE__. */ @@ -267,86 +275,27 @@ clib_mem_alloc_aligned_or_null (uword size, uword align) /* Alias to stack allocator for naming consistency. */ #define clib_mem_alloc_stack(bytes) __builtin_alloca(bytes) -always_inline uword -clib_mem_is_heap_object (void *p) -{ - void *heap = clib_mem_get_per_cpu_heap (); - int mspace_is_heap_object (void *msp, void *p); - - return mspace_is_heap_object (heap, p); -} - -always_inline void -clib_mem_free (void *p) -{ - u8 *heap = clib_mem_get_per_cpu_heap (); - - void mspace_put (void *msp, void *p_arg); - /* Make sure object is in the correct heap. */ - ASSERT (clib_mem_is_heap_object (p)); - - CLIB_MEM_POISON (p, clib_mem_size_nocheck (p)); - - mspace_put (heap, p); -} - -always_inline void * -clib_mem_realloc (void *p, uword new_size, uword old_size) -{ - /* By default use alloc, copy and free to emulate realloc. */ - void *q = clib_mem_alloc (new_size); - if (q) - { - uword copy_size; - if (old_size < new_size) - copy_size = old_size; - else - copy_size = new_size; - clib_memcpy_fast (q, p, copy_size); - clib_mem_free (p); - } - return q; -} - -always_inline uword -clib_mem_size (void *p) -{ - ASSERT (clib_mem_is_heap_object (p)); - return clib_mem_size_nocheck (p); -} - -always_inline void -clib_mem_free_s (void *p) -{ - uword size = clib_mem_size (p); - CLIB_MEM_UNPOISON (p, size); - memset_s_inline (p, size, 0, size); - clib_mem_free (p); -} - -always_inline void * +always_inline clib_mem_heap_t * clib_mem_get_heap (void) { return clib_mem_get_per_cpu_heap (); } -always_inline void * -clib_mem_set_heap (void *heap) +always_inline clib_mem_heap_t * +clib_mem_set_heap (clib_mem_heap_t * heap) { return clib_mem_set_per_cpu_heap (heap); } -void clib_mem_destroy_heap (void *heap); -void *clib_mem_create_heap (void *base, uword size, int is_locked, char *fmt, - ...); +void clib_mem_destroy_heap (clib_mem_heap_t * heap); +clib_mem_heap_t *clib_mem_create_heap (void *base, uword size, int is_locked, + char *fmt, ...); void clib_mem_main_init (); -void *clib_mem_init (void *heap, uword size); +void *clib_mem_init (void *base, uword size); void *clib_mem_init_with_page_size (uword memory_size, clib_mem_page_sz_t log2_page_sz); void *clib_mem_init_thread_safe (void *memory, uword memory_size); -void *clib_mem_init_thread_safe_numa (void *memory, uword memory_size, - u8 numa); void clib_mem_exit (void); @@ -379,14 +328,16 @@ typedef struct uword bytes_max; } clib_mem_usage_t; -void clib_mem_get_heap_usage (void *heap, clib_mem_usage_t * usage); +void clib_mem_get_heap_usage (clib_mem_heap_t * heap, + clib_mem_usage_t * usage); -void *clib_mem_get_heap_base (void *heap); -uword clib_mem_get_heap_size (void *heap); -uword clib_mem_get_heap_free_space (void *heap); +void *clib_mem_get_heap_base (clib_mem_heap_t * heap); +uword clib_mem_get_heap_size (clib_mem_heap_t * heap); +uword clib_mem_get_heap_free_space (clib_mem_heap_t * heap); u8 *format_clib_mem_usage (u8 * s, va_list * args); u8 *format_clib_mem_heap (u8 * s, va_list * va); +u8 *format_clib_mem_page_stats (u8 * s, va_list * va); /* Allocate virtual address space. */ always_inline void * @@ -403,7 +354,7 @@ clib_mem_vm_alloc (uword size) if (mmap_addr == (void *) -1) mmap_addr = 0; else - CLIB_MEM_UNPOISON (mmap_addr, size); + clib_mem_unpoison (mmap_addr, size); return mmap_addr; } @@ -427,36 +378,6 @@ int clib_mem_vm_unmap (void *base); clib_mem_vm_map_hdr_t *clib_mem_vm_get_next_map_hdr (clib_mem_vm_map_hdr_t * hdr); -typedef struct -{ -#define CLIB_MEM_VM_F_SHARED (1 << 0) -#define CLIB_MEM_VM_F_HUGETLB (1 << 1) -#define CLIB_MEM_VM_F_NUMA_PREFER (1 << 2) -#define CLIB_MEM_VM_F_NUMA_FORCE (1 << 3) -#define CLIB_MEM_VM_F_HUGETLB_PREALLOC (1 << 4) -#define CLIB_MEM_VM_F_LOCKED (1 << 5) - u32 flags; /**< vm allocation flags: -
CLIB_MEM_VM_F_SHARED: request shared memory, file - descriptor will be provided on successful allocation. -
CLIB_MEM_VM_F_HUGETLB: request hugepages. -
CLIB_MEM_VM_F_NUMA_PREFER: numa_node field contains valid - numa node preference. -
CLIB_MEM_VM_F_NUMA_FORCE: fail if setting numa policy fails. -
CLIB_MEM_VM_F_HUGETLB_PREALLOC: pre-allocate hugepages if - number of available pages is not sufficient. -
CLIB_MEM_VM_F_LOCKED: request locked memory. - */ - char *name; /**< Name for memory allocation, set by caller. */ - uword size; /**< Allocation size, set by caller. */ - int numa_node; /**< numa node preference. Valid if CLIB_MEM_VM_F_NUMA_PREFER set. */ - void *addr; /**< Pointer to allocated memory, set on successful allocation. */ - int fd; /**< File descriptor, set on successful allocation if CLIB_MEM_VM_F_SHARED is set. */ - int log2_page_size; /* Page size in log2 format, set on successful allocation. */ - int n_pages; /* Number of pages. */ - uword requested_va; /**< Request fixed position mapping */ -} clib_mem_vm_alloc_t; - - static_always_inline clib_mem_page_sz_t clib_mem_get_log2_page_size (void) { @@ -469,17 +390,26 @@ clib_mem_get_page_size (void) return 1ULL << clib_mem_main.log2_page_sz; } +static_always_inline void +clib_mem_set_log2_default_hugepage_size (clib_mem_page_sz_t log2_page_sz) +{ + clib_mem_main.log2_default_hugepage_sz = log2_page_sz; +} + static_always_inline clib_mem_page_sz_t clib_mem_get_log2_default_hugepage_size () { return clib_mem_main.log2_default_hugepage_sz; } +static_always_inline uword +clib_mem_get_default_hugepage_size (void) +{ + return 1ULL << clib_mem_main.log2_default_hugepage_sz; +} + int clib_mem_vm_create_fd (clib_mem_page_sz_t log2_page_size, char *fmt, ...); -clib_error_t *clib_mem_vm_ext_alloc (clib_mem_vm_alloc_t * a); -void clib_mem_vm_ext_free (clib_mem_vm_alloc_t * a); uword clib_mem_get_fd_page_size (int fd); -uword clib_mem_get_default_hugepage_size (void); clib_mem_page_sz_t clib_mem_get_fd_log2_page_size (int fd); uword clib_mem_vm_reserve (uword start, uword size, clib_mem_page_sz_t log2_page_sz); @@ -488,20 +418,9 @@ u64 *clib_mem_vm_get_paddr (void *mem, clib_mem_page_sz_t log2_page_size, void clib_mem_destroy (void); int clib_mem_set_numa_affinity (u8 numa_node, int force); int clib_mem_set_default_numa_affinity (); - -typedef struct -{ - uword size; /**< Map size */ - int fd; /**< File descriptor to be mapped */ - uword requested_va; /**< Request fixed position mapping */ - void *addr; /**< Pointer to mapped memory, if successful */ - u8 numa_node; -} clib_mem_vm_map_t; - -clib_error_t *clib_mem_vm_ext_map (clib_mem_vm_map_t * a); void clib_mem_vm_randomize_va (uword * requested_va, clib_mem_page_sz_t log2_page_size); -void mheap_trace (void *v, int enable); +void mheap_trace (clib_mem_heap_t * v, int enable); uword clib_mem_trace_enable_disable (uword enable); void clib_mem_trace (int enable); @@ -520,6 +439,8 @@ clib_mem_round_to_page_size (uword size, clib_mem_page_sz_t log2_page_size) typedef struct { + clib_mem_page_sz_t log2_page_sz; + uword total; uword mapped; uword not_mapped; uword per_numa[CLIB_MAX_NUMAS]; @@ -556,7 +477,7 @@ clib_mem_log2_page_size_validate (clib_mem_page_sz_t log2_page_size) static_always_inline uword clib_mem_page_bytes (clib_mem_page_sz_t log2_page_size) { - return 1 << clib_mem_log2_page_size_validate (log2_page_size); + return 1ULL << clib_mem_log2_page_size_validate (log2_page_size); } static_always_inline clib_error_t * @@ -565,6 +486,15 @@ clib_mem_get_last_error (void) return clib_mem_main.error; } +/* bulk allocator */ + +typedef void *clib_mem_bulk_handle_t; +clib_mem_bulk_handle_t clib_mem_bulk_init (u32 elt_sz, u32 align, + u32 min_elts_per_chunk); +void clib_mem_bulk_destroy (clib_mem_bulk_handle_t h); +void *clib_mem_bulk_alloc (clib_mem_bulk_handle_t h); +void clib_mem_bulk_free (clib_mem_bulk_handle_t h, void *p); +u8 *format_clib_mem_bulk (u8 *s, va_list *args); #include /* clib_panic */