X-Git-Url: https://gerrit.fd.io/r/gitweb?a=blobdiff_plain;f=src%2Fsvm%2Ffifo_segment.c;h=06b7f063136264014fcecfacb70a1e6a058e3bda;hb=0e6199dfa20d4c010bf3ac43e051f9f29c213478;hp=2db07593bffe0c52c370e335e270fadf4b438153;hpb=9e61d9a1aafbaecbc7b316fd94122b80c3f99905;p=vpp.git diff --git a/src/svm/fifo_segment.c b/src/svm/fifo_segment.c index 2db07593bff..06b7f063136 100644 --- a/src/svm/fifo_segment.c +++ b/src/svm/fifo_segment.c @@ -15,6 +15,12 @@ #include +static inline fifo_segment_slice_t * +fsh_slice_get (fifo_segment_header_t * fsh, u32 slice_index) +{ + return &fsh->slices[slice_index]; +} + static char *fifo_segment_mem_status_strings[] = { #define _(sym,str) str, foreach_segment_mem_status @@ -75,10 +81,44 @@ static inline uword fsh_n_cached_bytes (fifo_segment_header_t * fsh) { uword n_cached = clib_atomic_load_relax_n (&fsh->n_cached_bytes); - ASSERT (n_cached >= 0); return n_cached; } +static inline void +fsh_active_fifos_update (fifo_segment_header_t * fsh, int inc) +{ + clib_atomic_fetch_add_rel (&fsh->n_active_fifos, inc); +} + +static inline u32 +fsh_n_active_fifos (fifo_segment_header_t * fsh) +{ + return clib_atomic_load_relax_n (&fsh->n_active_fifos); +} + +static inline uword +fsh_virtual_mem (fifo_segment_header_t * fsh) +{ + fifo_segment_slice_t *fss; + uword total_vm = 0; + int i; + + for (i = 0; i < fsh->n_slices; i++) + { + fss = fsh_slice_get (fsh, i); + total_vm += clib_atomic_load_relax_n (&fss->virtual_mem); + } + return total_vm; +} + +void +fsh_virtual_mem_update (fifo_segment_header_t * fsh, u32 slice_index, + int n_bytes) +{ + fifo_segment_slice_t *fss = fsh_slice_get (fsh, slice_index); + fss->virtual_mem += n_bytes; +} + static void fsh_check_mem (fifo_segment_header_t * fsh) { @@ -96,18 +136,6 @@ fsh_check_mem (fifo_segment_header_t * fsh) fsh_update_free_bytes (fsh); } -static inline fifo_segment_slice_t * -fsh_slice_get (fifo_segment_header_t * fsh, u32 slice_index) -{ - return &fsh->slices[slice_index]; -} - -static inline void -fsh_active_fifos_update (fifo_segment_header_t * fsh, int inc) -{ - clib_atomic_fetch_add_rel (&fsh->n_active_fifos, inc); -} - /** * Initialize fifo segment shared header */ @@ -117,7 +145,7 @@ fifo_segment_init (fifo_segment_t * fs) fifo_segment_header_t *fsh; fifo_segment_slice_t *fss; ssvm_shared_header_t *sh; - u32 max_chunk_sz, max_chunks; + u32 max_chunk_sz; uword max_fifo; void *oldheap; int i; @@ -153,6 +181,7 @@ fifo_segment_init (fifo_segment_t * fs) { fss = fsh_slice_get (fsh, i); vec_validate_init_empty (fss->free_chunks, max_chunk_sz, 0); + vec_validate_init_empty (fss->num_chunks, max_chunk_sz, 0); clib_spinlock_init (&fss->chunk_lock); } @@ -160,8 +189,7 @@ fifo_segment_init (fifo_segment_t * fs) fsh->n_free_bytes = fsh_free_space (fsh); fsh->n_cached_bytes = 0; - max_chunks = fsh->n_free_bytes / FIFO_SEGMENT_MIN_FIFO_SIZE; - fsh->n_reserved_bytes = (max_chunks / 4) * sizeof (rb_node_t); + fsh->n_reserved_bytes = clib_min (0.01 * fsh->n_free_bytes, 256 << 10); sh->ready = 1; return (0); } @@ -290,7 +318,7 @@ fs_chunk_size_is_valid (fifo_segment_header_t * fsh, u32 size) * with a smaller FIFO. */ return size >= FIFO_SEGMENT_MIN_FIFO_SIZE - && size <= (1 << fsh->max_log2_chunk_size); + && size <= (1ULL << fsh->max_log2_chunk_size); } static svm_fifo_t * @@ -317,12 +345,75 @@ fs_try_alloc_fifo_freelist (fifo_segment_slice_t * fss, u32 fl_index) return f; } +svm_fifo_chunk_t * +fs_try_alloc_multi_chunk (fifo_segment_header_t * fsh, + fifo_segment_slice_t * fss, u32 data_bytes) +{ + u32 fl_index, fl_size, n_alloc = 0, req_bytes = data_bytes; + svm_fifo_chunk_t *c, *first = 0, *next; + + fl_index = fs_freelist_for_size (req_bytes); + if (fl_index > 0) + fl_index -= 1; + + fl_size = fs_freelist_index_to_size (fl_index); + + while (req_bytes) + { + c = fss->free_chunks[fl_index]; + if (c) + { + fss->free_chunks[fl_index] = c->next; + c->next = first; + first = c; + n_alloc += fl_size; + req_bytes -= clib_min (fl_size, req_bytes); + } + else + { + /* Failed to allocate with smaller chunks */ + if (fl_index == 0) + { + /* free all chunks if any allocated */ + c = first; + while (c) + { + fl_index = fs_freelist_for_size (c->length); + fl_size = fs_freelist_index_to_size (fl_index); + next = c->next; + c->next = fss->free_chunks[fl_index]; + fss->free_chunks[fl_index] = c; + fss->n_fl_chunk_bytes += fl_size; + c = next; + } + n_alloc = 0; + first = 0; + fl_index = fs_freelist_for_size (data_bytes); + if (fss->free_chunks[fl_index + 1]) + { + fl_index += 1; + fl_size = fs_freelist_index_to_size (fl_index); + continue; + } + + return 0; + } + fl_index -= 1; + fl_size = fl_size >> 1; + } + } + + fss->n_fl_chunk_bytes -= n_alloc; + fsh_cached_bytes_sub (fsh, n_alloc); + return first; +} + static svm_fifo_t * fs_try_alloc_fifo_freelist_multi_chunk (fifo_segment_header_t * fsh, fifo_segment_slice_t * fss, u32 data_bytes) { - svm_fifo_chunk_t *c, *first = 0, *last = 0; + svm_fifo_chunk_t *c, *first = 0, *last = 0, *next; u32 fl_index, fl_size, n_alloc = 0; svm_fifo_t *f; @@ -359,7 +450,7 @@ fs_try_alloc_fifo_freelist_multi_chunk (fifo_segment_header_t * fsh, c->next = first; first = c; n_alloc += fl_size; - data_bytes -= c->length; + data_bytes -= clib_min (fl_size, data_bytes); } else { @@ -372,11 +463,13 @@ fs_try_alloc_fifo_freelist_multi_chunk (fifo_segment_header_t * fsh, { fl_index = fs_freelist_for_size (c->length); fl_size = fs_freelist_index_to_size (fl_index); + next = c->next; c->next = fss->free_chunks[fl_index]; fss->free_chunks[fl_index] = c; fss->n_fl_chunk_bytes += fl_size; n_alloc -= fl_size; data_bytes += fl_size; + c = next; } first = last = 0; fl_index = fs_freelist_for_size (data_bytes); @@ -403,6 +496,52 @@ fs_try_alloc_fifo_freelist_multi_chunk (fifo_segment_header_t * fsh, return f; } +static int +fsh_try_alloc_chunk_batch (fifo_segment_header_t * fsh, + fifo_segment_slice_t * fss, + u32 fl_index, u32 batch_size) +{ + u32 rounded_data_size; + svm_fifo_chunk_t *c; + void *oldheap; + uword size; + u8 *cmem; + int i; + + rounded_data_size = fs_freelist_index_to_size (fl_index); + size = (uword) (sizeof (*c) + rounded_data_size) * batch_size; + + oldheap = ssvm_push_heap (fsh->ssvm_sh); + cmem = clib_mem_alloc_aligned_at_offset (size, CLIB_CACHE_LINE_BYTES, + 0 /* align_offset */ , + 0 /* os_out_of_memory */ ); + ssvm_pop_heap (oldheap); + + /* Out of space.. */ + if (cmem == 0) + return -1; + + /* Carve fifo + chunk space */ + for (i = 0; i < batch_size; i++) + { + c = (svm_fifo_chunk_t *) cmem; + c->start_byte = 0; + c->length = rounded_data_size; + c->enq_rb_index = RBTREE_TNIL_INDEX; + c->deq_rb_index = RBTREE_TNIL_INDEX; + c->next = fss->free_chunks[fl_index]; + fss->free_chunks[fl_index] = c; + cmem += sizeof (*c) + rounded_data_size; + } + + fss->num_chunks[fl_index] += batch_size; + fss->n_fl_chunk_bytes += batch_size * rounded_data_size; + fsh_cached_bytes_add (fsh, batch_size * rounded_data_size); + fsh_free_bytes_sub (fsh, size); + + return 0; +} + static int fs_try_alloc_fifo_batch (fifo_segment_header_t * fsh, fifo_segment_slice_t * fss, @@ -447,6 +586,7 @@ fs_try_alloc_fifo_batch (fifo_segment_header_t * fsh, fmem += hdrs + rounded_data_size; } + fss->num_chunks[fl_index] += batch_size; fss->n_fl_chunk_bytes += batch_size * rounded_data_size; fsh_cached_bytes_add (fsh, batch_size * rounded_data_size); fsh_free_bytes_sub (fsh, size); @@ -474,8 +614,11 @@ fs_try_alloc_fifo (fifo_segment_header_t * fsh, fifo_segment_slice_t * fss, min_size = clib_max ((fsh->pct_first_alloc * data_bytes) / 100, 4096); fl_index = fs_freelist_for_size (min_size); - fifo_sz = sizeof (svm_fifo_t) + sizeof (svm_fifo_chunk_t); - fifo_sz += 1 << max_log2 (min_size); + + if (fl_index >= vec_len (fss->free_chunks)) + return 0; + + clib_spinlock_lock (&fss->chunk_lock); if (fss->free_fifos && fss->free_chunks[fl_index]) { @@ -487,18 +630,28 @@ fs_try_alloc_fifo (fifo_segment_header_t * fsh, fifo_segment_slice_t * fss, } } - fsh_check_mem (fsh); + fifo_sz = sizeof (svm_fifo_t) + sizeof (svm_fifo_chunk_t); + fifo_sz += 1 << max_log2 (min_size); n_free_bytes = fsh_n_free_bytes (fsh); + if (fifo_sz * FIFO_SEGMENT_ALLOC_BATCH_SIZE < n_free_bytes) { - if (fs_try_alloc_fifo_batch (fsh, fss, fl_index, - FIFO_SEGMENT_ALLOC_BATCH_SIZE)) - goto done; - - f = fs_try_alloc_fifo_freelist (fss, fl_index); - if (f) - fsh_cached_bytes_sub (fsh, fs_freelist_index_to_size (fl_index)); - goto done; + if (!fs_try_alloc_fifo_batch (fsh, fss, fl_index, + FIFO_SEGMENT_ALLOC_BATCH_SIZE)) + { + f = fs_try_alloc_fifo_freelist (fss, fl_index); + if (f) + { + fsh_cached_bytes_sub (fsh, + fs_freelist_index_to_size (fl_index)); + goto done; + } + } + else + { + fsh_check_mem (fsh); + n_free_bytes = fsh_n_free_bytes (fsh); + } } if (fifo_sz <= n_free_bytes) { @@ -507,17 +660,25 @@ fs_try_alloc_fifo (fifo_segment_header_t * fsh, fifo_segment_slice_t * fss, ssvm_pop_heap (oldheap); if (f) { + clib_atomic_fetch_add_rel (&fss->num_chunks[fl_index], 1); fsh_free_bytes_sub (fsh, fifo_sz); goto done; } + fsh_check_mem (fsh); } - if (data_bytes <= fss->n_fl_chunk_bytes) - f = fs_try_alloc_fifo_freelist_multi_chunk (fsh, fss, data_bytes); + /* All failed, try to allocate min of data bytes and fifo sz */ + fifo_sz = clib_min (fifo_sz, data_bytes); + if (fifo_sz <= fss->n_fl_chunk_bytes) + f = fs_try_alloc_fifo_freelist_multi_chunk (fsh, fss, fifo_sz); done: + clib_spinlock_unlock (&fss->chunk_lock); if (f) - f->fs_hdr = fsh; + { + f->size = data_bytes; + f->fs_hdr = fsh; + } return f; } @@ -526,38 +687,74 @@ fsh_alloc_chunk (fifo_segment_header_t * fsh, u32 slice_index, u32 chunk_size) { fifo_segment_slice_t *fss; svm_fifo_chunk_t *c; - void *oldheap; int fl_index; fl_index = fs_freelist_for_size (chunk_size); fss = fsh_slice_get (fsh, slice_index); clib_spinlock_lock (&fss->chunk_lock); - c = fss->free_chunks[fl_index]; - if (!c) - { - fsh_check_mem (fsh); - chunk_size = fs_freelist_index_to_size (fl_index); - if (fsh_n_free_bytes (fsh) < chunk_size) - goto done; - - oldheap = ssvm_push_heap (fsh->ssvm_sh); - c = svm_fifo_chunk_alloc (chunk_size); - ssvm_pop_heap (oldheap); - - if (!c) - goto done; + ASSERT (vec_len (fss->free_chunks) > fl_index); + c = fss->free_chunks[fl_index]; - fsh_free_bytes_sub (fsh, chunk_size + sizeof (*c)); - } - else + if (c) { fss->free_chunks[fl_index] = c->next; c->next = 0; fss->n_fl_chunk_bytes -= fs_freelist_index_to_size (fl_index); fsh_cached_bytes_sub (fsh, fs_freelist_index_to_size (fl_index)); } + else + { + void *oldheap; + uword n_free; + u32 batch; + + chunk_size = fs_freelist_index_to_size (fl_index); + n_free = fsh_n_free_bytes (fsh); + + if (chunk_size <= n_free) + { + oldheap = ssvm_push_heap (fsh->ssvm_sh); + c = svm_fifo_chunk_alloc (chunk_size); + ssvm_pop_heap (oldheap); + + if (c) + { + clib_atomic_fetch_add_rel (&fss->num_chunks[fl_index], 1); + fsh_free_bytes_sub (fsh, chunk_size + sizeof (*c)); + goto done; + } + + fsh_check_mem (fsh); + n_free = fsh_n_free_bytes (fsh); + } + if (chunk_size <= fss->n_fl_chunk_bytes) + { + c = fs_try_alloc_multi_chunk (fsh, fss, chunk_size); + if (c) + goto done; + batch = n_free / FIFO_SEGMENT_MIN_FIFO_SIZE; + if (!batch || fsh_try_alloc_chunk_batch (fsh, fss, 0, batch)) + { + fsh_check_mem (fsh); + goto done; + } + } + if (chunk_size <= fss->n_fl_chunk_bytes + n_free) + { + u32 min_size = FIFO_SEGMENT_MIN_FIFO_SIZE; + + batch = (chunk_size - fss->n_fl_chunk_bytes) / min_size; + batch = clib_min (batch + 1, n_free / min_size); + if (fsh_try_alloc_chunk_batch (fsh, fss, 0, batch)) + { + fsh_check_mem (fsh); + goto done; + } + c = fs_try_alloc_multi_chunk (fsh, fss, chunk_size); + } + } done: @@ -603,6 +800,31 @@ fsh_collect_chunks (fifo_segment_header_t * fsh, u32 slice_index, fsh_slice_collect_chunks (fsh, fss, c); } +static inline void +fss_fifo_add_active_list (fifo_segment_slice_t * fss, svm_fifo_t * f) +{ + if (fss->fifos) + { + fss->fifos->prev = f; + f->next = fss->fifos; + } + fss->fifos = f; +} + +static inline void +fss_fifo_del_active_list (fifo_segment_slice_t * fss, svm_fifo_t * f) +{ + if (f->flags & SVM_FIFO_F_LL_TRACKED) + { + if (f->prev) + f->prev->next = f->next; + else + fss->fifos = f->next; + if (f->next) + f->next->prev = f->prev; + } +} + /** * Allocate fifo in fifo segment */ @@ -616,6 +838,9 @@ fifo_segment_alloc_fifo_w_slice (fifo_segment_t * fs, u32 slice_index, ASSERT (slice_index < fs->n_slices); + if (PREDICT_FALSE (data_bytes > 1 << fsh->max_log2_chunk_size)) + return 0; + fss = fsh_slice_get (fsh, slice_index); f = fs_try_alloc_fifo (fsh, fss, data_bytes); if (!f) @@ -631,12 +856,7 @@ fifo_segment_alloc_fifo_w_slice (fifo_segment_t * fs, u32 slice_index, * only one. */ if (ftype == FIFO_SEGMENT_RX_FIFO) { - if (fss->fifos) - { - fss->fifos->prev = f; - f->next = fss->fifos; - } - fss->fifos = f; + fss_fifo_add_active_list (fss, f); f->flags |= SVM_FIFO_F_LL_TRACKED; svm_fifo_init_ooo_lookup (f, 0 /* ooo enq */ ); @@ -647,6 +867,7 @@ fifo_segment_alloc_fifo_w_slice (fifo_segment_t * fs, u32 slice_index, } fsh_active_fifos_update (fsh, 1); + fss->virtual_mem += svm_fifo_size (f); done: return (f); @@ -671,20 +892,10 @@ fifo_segment_free_fifo (fifo_segment_t * fs, svm_fifo_t * f) /* Remove from active list. Only rx fifos are tracked */ if (f->flags & SVM_FIFO_F_LL_TRACKED) { - if (f->prev) - f->prev->next = f->next; - else - fss->fifos = f->next; - if (f->next) - f->next->prev = f->prev; + fss_fifo_del_active_list (fss, f); f->flags &= ~SVM_FIFO_F_LL_TRACKED; } - /* Add to free list */ - f->next = fss->free_fifos; - f->prev = 0; - fss->free_fifos = f; - /* Free fifo chunks */ fsh_slice_collect_chunks (fsh, fss, f->start_chunk); @@ -701,9 +912,62 @@ fifo_segment_free_fifo (fifo_segment_t * fs, svm_fifo_t * f) f->master_thread_index = ~0; } + fss->virtual_mem -= svm_fifo_size (f); + + /* Add to free list */ + f->next = fss->free_fifos; + f->prev = 0; + fss->free_fifos = f; + fsh_active_fifos_update (fsh, -1); } +void +fifo_segment_detach_fifo (fifo_segment_t * fs, svm_fifo_t * f) +{ + fifo_segment_slice_t *fss; + svm_fifo_chunk_t *c; + u32 fl_index; + + ASSERT (f->refcnt == 1); + + fss = fsh_slice_get (fs->h, f->slice_index); + fss->virtual_mem -= svm_fifo_size (f); + if (f->flags & SVM_FIFO_F_LL_TRACKED) + fss_fifo_del_active_list (fss, f); + + c = f->start_chunk; + while (c) + { + fl_index = fs_freelist_for_size (c->length); + clib_atomic_fetch_sub_rel (&fss->num_chunks[fl_index], 1); + c = c->next; + } +} + +void +fifo_segment_attach_fifo (fifo_segment_t * fs, svm_fifo_t * f, + u32 slice_index) +{ + fifo_segment_slice_t *fss; + svm_fifo_chunk_t *c; + u32 fl_index; + + f->slice_index = slice_index; + fss = fsh_slice_get (fs->h, f->slice_index); + fss->virtual_mem += svm_fifo_size (f); + if (f->flags & SVM_FIFO_F_LL_TRACKED) + fss_fifo_add_active_list (fss, f); + + c = f->start_chunk; + while (c) + { + fl_index = fs_freelist_for_size (c->length); + clib_atomic_fetch_add_rel (&fss->num_chunks[fl_index], 1); + c = c->next; + } +} + int fifo_segment_prealloc_fifo_hdrs (fifo_segment_t * fs, u32 slice_index, u32 batch_size) @@ -791,6 +1055,7 @@ fifo_segment_prealloc_fifo_chunks (fifo_segment_t * fs, u32 slice_index, fsh_cached_bytes_add (fsh, rounded_data_size); } + fss->num_chunks[fl_index] += batch_size; fss->n_fl_chunk_bytes += batch_size * rounded_data_size; fsh_free_bytes_sub (fsh, size); @@ -866,7 +1131,7 @@ fifo_segment_preallocate_fifo_pairs (fifo_segment_t * fs, u32 fifo_segment_num_fifos (fifo_segment_t * fs) { - return clib_atomic_load_relax_n (&fs->h->n_active_fifos); + return fsh_n_active_fifos (fs->h); } static u32 @@ -998,6 +1263,12 @@ fifo_segment_cached_bytes (fifo_segment_t * fs) return fsh_n_cached_bytes (fs->h); } +uword +fifo_segment_available_bytes (fifo_segment_t * fs) +{ + return fsh_n_free_bytes (fs->h) + fsh_n_cached_bytes (fs->h); +} + uword fifo_segment_fl_chunk_bytes (fifo_segment_t * fs) { @@ -1018,17 +1289,7 @@ fifo_segment_fl_chunk_bytes (fifo_segment_t * fs) u8 fifo_segment_has_fifos (fifo_segment_t * fs) { - fifo_segment_header_t *fsh = fs->h; - fifo_segment_slice_t *fss; - int slice_index; - - for (slice_index = 0; slice_index < fs->n_slices; slice_index++) - { - fss = fsh_slice_get (fsh, slice_index); - if (fss->fifos) - return 1; - } - return 0; + return (fsh_n_active_fifos (fs->h) != 0); } svm_fifo_t * @@ -1111,12 +1372,13 @@ format_fifo_segment_type (u8 * s, va_list * args) u8 * format_fifo_segment (u8 * s, va_list * args) { - u32 count, indent, active_fifos, free_fifos, fifo_hdr = 0; + u32 count, indent, active_fifos, free_fifos; fifo_segment_t *fs = va_arg (*args, fifo_segment_t *); int verbose __attribute__ ((unused)) = va_arg (*args, int); uword est_chunk_bytes, est_free_seg_bytes, free_chunks; uword chunk_bytes = 0, free_seg_bytes, chunk_size; uword tracked_cached_bytes; + uword fifo_hdr = 0, reserved; fifo_segment_header_t *fsh; fifo_segment_slice_t *fss; svm_fifo_chunk_t *c; @@ -1124,7 +1386,7 @@ format_fifo_segment (u8 * s, va_list * args) char *address; size_t size; int i; - uword allocated, in_use; + uword allocated, in_use, virt; f64 usage; fifo_segment_mem_status_t mem_st; @@ -1152,8 +1414,9 @@ format_fifo_segment (u8 * s, va_list * args) free_chunks = fifo_segment_num_free_chunks (fs, ~0); if (free_chunks) - s = format (s, "\n\n%UFree chunks by size:\n", format_white_space, - indent + 2); + s = + format (s, "\n\n%UFree/Allocated chunks by size:\n", format_white_space, + indent + 2); else s = format (s, "\n"); @@ -1163,7 +1426,7 @@ format_fifo_segment (u8 * s, va_list * args) for (i = 0; i < vec_len (fss->free_chunks); i++) { c = fss->free_chunks[i]; - if (c == 0) + if (c == 0 && fss->num_chunks[i] == 0) continue; count = 0; while (c) @@ -1173,8 +1436,8 @@ format_fifo_segment (u8 * s, va_list * args) } chunk_size = fs_freelist_index_to_size (i); - s = format (s, "%U%-5u kB: %u\n", format_white_space, indent + 2, - chunk_size >> 10, count); + s = format (s, "%U%-5u kB: %u/%u\n", format_white_space, indent + 2, + chunk_size >> 10, count, fss->num_chunks[i]); chunk_bytes += count * chunk_size; } @@ -1190,26 +1453,26 @@ format_fifo_segment (u8 * s, va_list * args) in_use = fifo_segment_size (fs) - est_free_seg_bytes - tracked_cached_bytes; usage = (100.0 * in_use) / allocated; mem_st = fifo_segment_get_mem_status (fs); - - s = format (s, "\n%Useg free bytes: %U (%lu) estimated: %U (%lu)\n", - format_white_space, indent + 2, format_memory_size, - free_seg_bytes, free_seg_bytes, format_memory_size, - est_free_seg_bytes, est_free_seg_bytes); - s = - format (s, - "%Uchunk free bytes: %U (%lu) estimated: %U (%lu) tracked: %U (%lu)\n", - format_white_space, indent + 2, format_memory_size, chunk_bytes, - chunk_bytes, format_memory_size, est_chunk_bytes, est_chunk_bytes, - format_memory_size, tracked_cached_bytes, tracked_cached_bytes); - s = - format (s, "%Ufifo hdr free bytes: %U (%u) reserved %U (%lu)\n", - format_white_space, indent + 2, format_memory_size, fifo_hdr, - fifo_hdr, format_memory_size, fsh->n_reserved_bytes, - fsh->n_reserved_bytes); - s = - format (s, "%Usegment usage: %.2f%% (%U / %U) %s\n", format_white_space, - indent + 2, usage, format_memory_size, in_use, format_memory_size, - allocated, fifo_segment_mem_status_strings[mem_st]); + virt = fsh_virtual_mem (fsh); + reserved = fsh->n_reserved_bytes; + + s = format (s, "\n%Useg free bytes: %U (%lu) estimated: %U (%lu) reserved:" + " %U (%lu)\n", format_white_space, indent + 2, + format_memory_size, free_seg_bytes, free_seg_bytes, + format_memory_size, est_free_seg_bytes, est_free_seg_bytes, + format_memory_size, reserved, reserved); + s = format (s, "%Uchunk free bytes: %U (%lu) estimated: %U (%lu) tracked:" + " %U (%lu)\n", format_white_space, indent + 2, + format_memory_size, chunk_bytes, chunk_bytes, + format_memory_size, est_chunk_bytes, est_chunk_bytes, + format_memory_size, tracked_cached_bytes, tracked_cached_bytes); + s = format (s, "%Ufifo active: %u hdr free bytes: %U (%u) \n", + format_white_space, indent + 2, fsh->n_active_fifos, + format_memory_size, fifo_hdr, fifo_hdr); + s = format (s, "%Usegment usage: %.2f%% (%U / %U) virt: %U status: %s\n", + format_white_space, indent + 2, usage, format_memory_size, + in_use, format_memory_size, allocated, format_memory_size, virt, + fifo_segment_mem_status_strings[mem_st]); s = format (s, "\n"); return s;