+svm_fifo_chunk_t *
+fsh_alloc_chunk (fifo_segment_header_t * fsh, u32 slice_index, u32 chunk_size)
+{
+ fifo_segment_slice_t *fss;
+ svm_fifo_chunk_t *c;
+ void *oldheap;
+ int fl_index;
+ uword n_free;
+
+ fl_index = fs_freelist_for_size (chunk_size);
+ fss = fsh_slice_get (fsh, slice_index);
+
+ clib_spinlock_lock (&fss->chunk_lock);
+
+ c = fss->free_chunks[fl_index];
+
+ if (c)
+ {
+ fss->free_chunks[fl_index] = c->next;
+ c->next = 0;
+ fss->n_fl_chunk_bytes -= fs_freelist_index_to_size (fl_index);
+ fsh_cached_bytes_sub (fsh, fs_freelist_index_to_size (fl_index));
+ }
+ else if (chunk_size <= (n_free = fsh_n_free_bytes (fsh)))
+ {
+ fsh_check_mem (fsh);
+
+ chunk_size = fs_freelist_index_to_size (fl_index);
+ if (n_free < chunk_size)
+ goto done;
+
+ oldheap = ssvm_push_heap (fsh->ssvm_sh);
+ c = svm_fifo_chunk_alloc (chunk_size);
+ ssvm_pop_heap (oldheap);
+
+ if (!c)
+ goto done;
+
+ fsh_free_bytes_sub (fsh, chunk_size + sizeof (*c));
+ }
+ else if (chunk_size <= fss->n_fl_chunk_bytes)
+ {
+ c = fs_try_alloc_multi_chunk (fsh, fss, chunk_size);
+ }
+ else if (chunk_size <= fss->n_fl_chunk_bytes + n_free)
+ {
+ u32 min_size = FIFO_SEGMENT_MIN_FIFO_SIZE;
+ u32 batch;
+
+ fsh_check_mem (fsh);
+ batch = (chunk_size - fss->n_fl_chunk_bytes) / min_size;
+ batch = clib_min (batch + 1, n_free / min_size);
+ if (!fsh_try_alloc_chunk_batch (fsh, fss, 0, batch))
+ c = fs_try_alloc_multi_chunk (fsh, fss, chunk_size);
+ }
+
+done:
+
+ clib_spinlock_unlock (&fss->chunk_lock);
+
+ return c;
+}
+
+static void
+fsh_slice_collect_chunks (fifo_segment_header_t * fsh,
+ fifo_segment_slice_t * fss, svm_fifo_chunk_t * c)
+{
+ svm_fifo_chunk_t *next;
+ int fl_index;
+ u32 n_collect = 0;
+
+ clib_spinlock_lock (&fss->chunk_lock);
+
+ while (c)
+ {
+ next = c->next;
+ fl_index = fs_freelist_for_size (c->length);
+ c->next = fss->free_chunks[fl_index];
+ c->enq_rb_index = RBTREE_TNIL_INDEX;
+ c->deq_rb_index = RBTREE_TNIL_INDEX;
+ fss->free_chunks[fl_index] = c;
+ n_collect += fs_freelist_index_to_size (fl_index);
+ c = next;
+ }
+
+ fss->n_fl_chunk_bytes += n_collect;
+ fsh_cached_bytes_add (fsh, n_collect);
+
+ clib_spinlock_unlock (&fss->chunk_lock);
+}
+
+void
+fsh_collect_chunks (fifo_segment_header_t * fsh, u32 slice_index,
+ svm_fifo_chunk_t * c)
+{
+ fifo_segment_slice_t *fss;
+ fss = fsh_slice_get (fsh, slice_index);
+ fsh_slice_collect_chunks (fsh, fss, c);
+}
+