X-Git-Url: https://gerrit.fd.io/r/gitweb?a=blobdiff_plain;f=src%2Fvppinfra%2Fbihash_template.h;h=d6aa3c9f6eff4c3bf534739a1fa133935041dddf;hb=7be474635dad869975b3a5dbcaaf17d14f2fb104;hp=98dcf14673f8db97d0978cde5021329a9b6c8d78;hpb=a713254d689deee1718bb515c3961bf4ed435504;p=vpp.git diff --git a/src/vppinfra/bihash_template.h b/src/vppinfra/bihash_template.h index 98dcf14673f..d6aa3c9f6ef 100644 --- a/src/vppinfra/bihash_template.h +++ b/src/vppinfra/bihash_template.h @@ -44,6 +44,11 @@ #define BIHASH_FREELIST_LENGTH 17 #endif +/* default is 2MB, use 30 for 1GB */ +#ifndef BIHASH_LOG2_HUGEPAGE_SIZE +#define BIHASH_LOG2_HUGEPAGE_SIZE 21 +#endif + #define _bv(a,b) a##b #define __bv(a,b) _bv(a,b) #define BV(a) __bv(a,BIHASH_TYPE) @@ -52,6 +57,10 @@ #define __bvt(a,b) _bvt(a,b) #define BVT(a) __bvt(a,BIHASH_TYPE) +#define _bvs(a,b) struct a##b +#define __bvs(a,b) _bvs(a,b) +#define BVS(a) __bvs(a,BIHASH_TYPE) + #if _LP64 == 0 #define OVERFLOW_ASSERT(x) ASSERT(((x) & 0xFFFFFFFF00000000ULL) == 0) #define u64_to_pointer(x) (void *)(u32)((x)) @@ -99,6 +108,7 @@ typedef CLIB_PACKED (struct { */ u64 alloc_arena_next; /* Next offset from alloc_arena to allocate, definitely NOT a constant */ u64 alloc_arena_size; /* Size of the arena */ + u64 alloc_arena_mapped; /* Size of the mapped memory in the arena */ /* Two SVM pointers stored as 8-byte integers */ u64 alloc_lock_as_u64; u64 buckets_as_u64; @@ -107,63 +117,136 @@ typedef CLIB_PACKED (struct { u32 nbuckets; /* Number of buckets */ /* Set when header valid */ volatile u32 ready; - u64 pad[2]; + u64 pad[1]; }) BVT (clib_bihash_shared_header); /* *INDENT-ON* */ STATIC_ASSERT_SIZEOF (BVT (clib_bihash_shared_header), 8 * sizeof (u64)); -typedef struct +typedef +BVS (clib_bihash) { BVT (clib_bihash_bucket) * buckets; volatile u32 *alloc_lock; - BVT (clib_bihash_value) ** working_copies; + BVT (clib_bihash_value) ** working_copies; int *working_copy_lengths; - BVT (clib_bihash_bucket) saved_bucket; + BVT (clib_bihash_bucket) saved_bucket; u32 nbuckets; u32 log2_nbuckets; + u64 memory_size; u8 *name; u64 *freelists; #if BIHASH_32_64_SVM - BVT (clib_bihash_shared_header) * sh; + BVT (clib_bihash_shared_header) * sh; int memfd; #else - BVT (clib_bihash_shared_header) sh; + BVT (clib_bihash_shared_header) sh; #endif u64 alloc_arena; /* Base of the allocation arena */ + volatile u8 instantiated; /** * A custom format function to print the Key and Value of bihash_key instead of default hexdump */ format_function_t *fmt_fn; + /** Optional statistics-gathering callback */ +#if BIHASH_ENABLE_STATS + void (*inc_stats_callback) (BVS (clib_bihash) *, int stat_id, u64 count); + + /** Statistics callback context (e.g. address of stats data structure) */ + void *inc_stats_context; +#endif + } BVT (clib_bihash); +typedef struct +{ + BVT (clib_bihash) * h; + char *name; + u32 nbuckets; + uword memory_size; + format_function_t *fmt_fn; + u8 instantiate_immediately; + u8 dont_add_to_all_bihash_list; +} BVT (clib_bihash_init2_args); + +extern void **clib_all_bihashes; + #if BIHASH_32_64_SVM #undef alloc_arena_next #undef alloc_arena_size +#undef alloc_arena_mapped #undef alloc_arena #undef CLIB_BIHASH_READY_MAGIC #define alloc_arena_next(h) (((h)->sh)->alloc_arena_next) #define alloc_arena_size(h) (((h)->sh)->alloc_arena_size) +#define alloc_arena_mapped(h) (((h)->sh)->alloc_arena_mapped) #define alloc_arena(h) ((h)->alloc_arena) #define CLIB_BIHASH_READY_MAGIC 0xFEEDFACE #else #undef alloc_arena_next #undef alloc_arena_size +#undef alloc_arena_mapped #undef alloc_arena #undef CLIB_BIHASH_READY_MAGIC #define alloc_arena_next(h) ((h)->sh.alloc_arena_next) #define alloc_arena_size(h) ((h)->sh.alloc_arena_size) +#define alloc_arena_mapped(h) ((h)->sh.alloc_arena_mapped) #define alloc_arena(h) ((h)->alloc_arena) #define CLIB_BIHASH_READY_MAGIC 0 #endif +#ifndef BIHASH_STAT_IDS +#define BIHASH_STAT_IDS 1 + +#define foreach_bihash_stat \ +_(alloc_add) \ +_(add) \ +_(split_add) \ +_(replace) \ +_(update) \ +_(del) \ +_(del_free) \ +_(linear) \ +_(resplit) \ +_(working_copy_lost) \ +_(splits) /* must be last */ + +typedef enum +{ +#define _(a) BIHASH_STAT_##a, + foreach_bihash_stat +#undef _ + BIHASH_STAT_N_STATS, +} BVT (clib_bihash_stat_id); +#endif /* BIHASH_STAT_IDS */ + +static inline void BV (clib_bihash_increment_stat) (BVT (clib_bihash) * h, + int stat_id, u64 count) +{ +#if BIHASH_ENABLE_STATS + if (PREDICT_FALSE (h->inc_stats_callback != 0)) + h->inc_stats_callback (h, stat_id, count); +#endif +} + +#if BIHASH_ENABLE_STATS +static inline void BV (clib_bihash_set_stats_callback) + (BVT (clib_bihash) * h, void (*cb) (BVT (clib_bihash) *, int, u64), + void *ctx) +{ + h->inc_stats_callback = cb; + h->inc_stats_context = ctx; +} +#endif + + static inline void BV (clib_bihash_alloc_lock) (BVT (clib_bihash) * h) { while (__atomic_test_and_set (h->alloc_lock, __ATOMIC_ACQUIRE)) @@ -177,25 +260,25 @@ static inline void BV (clib_bihash_alloc_unlock) (BVT (clib_bihash) * h) static inline void BV (clib_bihash_lock_bucket) (BVT (clib_bihash_bucket) * b) { - BVT (clib_bihash_bucket) unlocked_bucket, locked_bucket; + /* *INDENT-OFF* */ + BVT (clib_bihash_bucket) mask = { .lock = 1 }; + /* *INDENT-ON* */ + u64 old; - do +try_again: + old = clib_atomic_fetch_or (&b->as_u64, mask.as_u64); + + if (PREDICT_FALSE (old & mask.as_u64)) { - locked_bucket.as_u64 = unlocked_bucket.as_u64 = b->as_u64; - unlocked_bucket.lock = 0; - locked_bucket.lock = 1; + /* somebody else flipped the bit, try again */ CLIB_PAUSE (); + goto try_again; } - while (__atomic_compare_exchange_n (&b->as_u64, &unlocked_bucket.as_u64, - locked_bucket.as_u64, 1 /* weak */ , - __ATOMIC_ACQUIRE, - __ATOMIC_ACQUIRE) == 0); } static inline void BV (clib_bihash_unlock_bucket) (BVT (clib_bihash_bucket) * b) { - CLIB_MEMORY_BARRIER (); b->lock = 0; } @@ -212,7 +295,10 @@ static inline int BV (clib_bihash_bucket_is_empty) (BVT (clib_bihash_bucket) * b) { /* Note: applied to locked buckets, test offset */ - return b->offset == 0; + if (BIHASH_KVP_AT_BUCKET_LEVEL == 0) + return b->offset == 0; + else + return (b->log2_pages == 0 && b->refcnt == 1); } static inline uword BV (clib_bihash_get_offset) (BVT (clib_bihash) * h, @@ -229,10 +315,12 @@ static inline uword BV (clib_bihash_get_offset) (BVT (clib_bihash) * h, void BV (clib_bihash_init) (BVT (clib_bihash) * h, char *name, u32 nbuckets, uword memory_size); +void BV (clib_bihash_init2) (BVT (clib_bihash_init2_args) * a); + #if BIHASH_32_64_SVM -void BV (clib_bihash_master_init_svm) +void BV (clib_bihash_initiator_init_svm) (BVT (clib_bihash) * h, char *name, u32 nbuckets, u64 memory_size); -void BV (clib_bihash_slave_init_svm) +void BV (clib_bihash_responder_init_svm) (BVT (clib_bihash) * h, char *name, int fd); #endif @@ -253,23 +341,58 @@ int BV (clib_bihash_search) (BVT (clib_bihash) * h, BVT (clib_bihash_kv) * search_v, BVT (clib_bihash_kv) * return_v); +#define BIHASH_WALK_STOP 0 +#define BIHASH_WALK_CONTINUE 1 + +typedef + int (*BV (clib_bihash_foreach_key_value_pair_cb)) (BVT (clib_bihash_kv) *, + void *); void BV (clib_bihash_foreach_key_value_pair) (BVT (clib_bihash) * h, - void *callback, void *arg); + BV + (clib_bihash_foreach_key_value_pair_cb) + cb, void *arg); +void *clib_all_bihash_set_heap (void); +void clib_bihash_copied (void *dst, void *src); format_function_t BV (format_bihash); format_function_t BV (format_bihash_kvp); format_function_t BV (format_bihash_lru); +static inline +BVT (clib_bihash_bucket) * +BV (clib_bihash_get_bucket) (BVT (clib_bihash) * h, u64 hash) +{ +#if BIHASH_KVP_AT_BUCKET_LEVEL + uword offset; + offset = (hash & (h->nbuckets - 1)); + offset = offset * (sizeof (BVT (clib_bihash_bucket)) + + (BIHASH_KVP_PER_PAGE * sizeof (BVT (clib_bihash_kv)))); + return ((BVT (clib_bihash_bucket) *) (((u8 *) h->buckets) + offset)); +#else + return h->buckets + (hash & (h->nbuckets - 1)); +#endif +} + static inline int BV (clib_bihash_search_inline_with_hash) (BVT (clib_bihash) * h, u64 hash, BVT (clib_bihash_kv) * key_result) { - u32 bucket_index; BVT (clib_bihash_value) * v; BVT (clib_bihash_bucket) * b; int i, limit; - bucket_index = hash & (h->nbuckets - 1); - b = &h->buckets[bucket_index]; + /* *INDENT-OFF* */ + static const BVT (clib_bihash_bucket) mask = { + .linear_search = 1, + .log2_pages = -1 + }; + /* *INDENT-ON* */ + +#if BIHASH_LAZY_INSTANTIATE + if (PREDICT_FALSE (h->instantiated == 0)) + return -1; +#endif + + b = BV (clib_bihash_get_bucket) (h, hash); if (PREDICT_FALSE (BV (clib_bihash_bucket_is_empty) (b))) return -1; @@ -281,15 +404,18 @@ static inline int BV (clib_bihash_search_inline_with_hash) CLIB_PAUSE (); } - hash >>= h->log2_nbuckets; - v = BV (clib_bihash_get_value) (h, b->offset); /* If the bucket has unresolvable collisions, use linear search */ limit = BIHASH_KVP_PER_PAGE; - v += (b->linear_search == 0) ? hash & ((1 << b->log2_pages) - 1) : 0; - if (PREDICT_FALSE (b->linear_search)) - limit <<= b->log2_pages; + + if (PREDICT_FALSE (b->as_u64 & mask.as_u64)) + { + if (PREDICT_FALSE (b->linear_search)) + limit <<= b->log2_pages; + else + v += extract_bits (hash, h->log2_nbuckets, b->log2_pages); + } for (i = 0; i < limit; i++) { @@ -315,49 +441,59 @@ static inline int BV (clib_bihash_search_inline) static inline void BV (clib_bihash_prefetch_bucket) (BVT (clib_bihash) * h, u64 hash) { - u32 bucket_index; - BVT (clib_bihash_bucket) * b; - - bucket_index = hash & (h->nbuckets - 1); - b = &h->buckets[bucket_index]; - - CLIB_PREFETCH (b, CLIB_CACHE_LINE_BYTES, READ); + CLIB_PREFETCH (BV (clib_bihash_get_bucket) (h, hash), + BIHASH_BUCKET_PREFETCH_CACHE_LINES * CLIB_CACHE_LINE_BYTES, + LOAD); } static inline void BV (clib_bihash_prefetch_data) (BVT (clib_bihash) * h, u64 hash) { - u32 bucket_index; BVT (clib_bihash_value) * v; BVT (clib_bihash_bucket) * b; - bucket_index = hash & (h->nbuckets - 1); - b = &h->buckets[bucket_index]; +#if BIHASH_LAZY_INSTANTIATE + if (PREDICT_FALSE (h->instantiated == 0)) + return; +#endif + + b = BV (clib_bihash_get_bucket) (h, hash); if (PREDICT_FALSE (BV (clib_bihash_bucket_is_empty) (b))) return; - hash >>= h->log2_nbuckets; v = BV (clib_bihash_get_value) (h, b->offset); - v += (b->linear_search == 0) ? hash & ((1 << b->log2_pages) - 1) : 0; + if (PREDICT_FALSE (b->log2_pages && b->linear_search == 0)) + v += extract_bits (hash, h->log2_nbuckets, b->log2_pages); - CLIB_PREFETCH (v, CLIB_CACHE_LINE_BYTES, READ); + CLIB_PREFETCH (v, BIHASH_KVP_PER_PAGE * sizeof (BVT (clib_bihash_kv)), + LOAD); } static inline int BV (clib_bihash_search_inline_2_with_hash) (BVT (clib_bihash) * h, u64 hash, BVT (clib_bihash_kv) * search_key, BVT (clib_bihash_kv) * valuep) { - u32 bucket_index; BVT (clib_bihash_value) * v; BVT (clib_bihash_bucket) * b; int i, limit; +/* *INDENT-OFF* */ + static const BVT (clib_bihash_bucket) mask = { + .linear_search = 1, + .log2_pages = -1 + }; +/* *INDENT-ON* */ + ASSERT (valuep); - bucket_index = hash & (h->nbuckets - 1); - b = &h->buckets[bucket_index]; +#if BIHASH_LAZY_INSTANTIATE + if (PREDICT_FALSE (h->instantiated == 0)) + return -1; +#endif + + b = BV (clib_bihash_get_bucket) (h, hash); if (PREDICT_FALSE (BV (clib_bihash_bucket_is_empty) (b))) return -1; @@ -369,14 +505,18 @@ static inline int BV (clib_bihash_search_inline_2_with_hash) CLIB_PAUSE (); } - hash >>= h->log2_nbuckets; v = BV (clib_bihash_get_value) (h, b->offset); /* If the bucket has unresolvable collisions, use linear search */ limit = BIHASH_KVP_PER_PAGE; - v += (b->linear_search == 0) ? hash & ((1 << b->log2_pages) - 1) : 0; - if (PREDICT_FALSE (b->linear_search)) - limit <<= b->log2_pages; + + if (PREDICT_FALSE (b->as_u64 & mask.as_u64)) + { + if (PREDICT_FALSE (b->linear_search)) + limit <<= b->log2_pages; + else + v += extract_bits (hash, h->log2_nbuckets, b->log2_pages); + } for (i = 0; i < limit; i++) {