+svm_fifo_chunk_t *
+fsh_alloc_chunk (fifo_segment_header_t * fsh, u32 slice_index, u32 chunk_size)
+{
+ fifo_segment_slice_t *fss;
+ svm_fifo_chunk_t *c;
+ int fl_index;
+
+ fl_index = fs_freelist_for_size (chunk_size);
+ fss = fsh_slice_get (fsh, slice_index);
+
+ clib_spinlock_lock (&fss->chunk_lock);
+
+ c = fss->free_chunks[fl_index];
+
+ if (c)
+ {
+ fss->free_chunks[fl_index] = c->next;
+ c->next = 0;
+ fss->n_fl_chunk_bytes -= fs_freelist_index_to_size (fl_index);
+ fsh_cached_bytes_sub (fsh, fs_freelist_index_to_size (fl_index));
+ }
+ else
+ {
+ void *oldheap;
+ uword n_free;
+ u32 batch;
+
+ chunk_size = fs_freelist_index_to_size (fl_index);
+ n_free = fsh_n_free_bytes (fsh);
+
+ if (chunk_size <= n_free)
+ {
+ oldheap = ssvm_push_heap (fsh->ssvm_sh);
+ c = svm_fifo_chunk_alloc (chunk_size);
+ ssvm_pop_heap (oldheap);
+
+ if (c)
+ {
+ clib_atomic_fetch_add_rel (&fss->num_chunks[fl_index], 1);
+ fsh_free_bytes_sub (fsh, chunk_size + sizeof (*c));
+ goto done;
+ }
+
+ fsh_check_mem (fsh);
+ n_free = fsh_n_free_bytes (fsh);
+ }
+ if (chunk_size <= fss->n_fl_chunk_bytes)
+ {
+ c = fs_try_alloc_multi_chunk (fsh, fss, chunk_size);
+ if (c)
+ goto done;
+ batch = n_free / FIFO_SEGMENT_MIN_FIFO_SIZE;
+ if (!batch || fsh_try_alloc_chunk_batch (fsh, fss, 0, batch))
+ {
+ fsh_check_mem (fsh);
+ goto done;
+ }
+ }
+ if (chunk_size <= fss->n_fl_chunk_bytes + n_free)
+ {
+ u32 min_size = FIFO_SEGMENT_MIN_FIFO_SIZE;
+
+ batch = (chunk_size - fss->n_fl_chunk_bytes) / min_size;
+ batch = clib_min (batch + 1, n_free / min_size);
+ if (fsh_try_alloc_chunk_batch (fsh, fss, 0, batch))
+ {
+ fsh_check_mem (fsh);
+ goto done;
+ }
+ c = fs_try_alloc_multi_chunk (fsh, fss, chunk_size);
+ }
+ }
+
+done:
+
+ clib_spinlock_unlock (&fss->chunk_lock);
+
+ return c;
+}
+
+static void
+fsh_slice_collect_chunks (fifo_segment_header_t * fsh,
+ fifo_segment_slice_t * fss, svm_fifo_chunk_t * c)
+{
+ svm_fifo_chunk_t *next;
+ int fl_index;
+ u32 n_collect = 0;
+
+ clib_spinlock_lock (&fss->chunk_lock);
+
+ while (c)
+ {
+ next = c->next;
+ fl_index = fs_freelist_for_size (c->length);
+ c->next = fss->free_chunks[fl_index];
+ c->enq_rb_index = RBTREE_TNIL_INDEX;
+ c->deq_rb_index = RBTREE_TNIL_INDEX;
+ fss->free_chunks[fl_index] = c;
+ n_collect += fs_freelist_index_to_size (fl_index);
+ c = next;
+ }
+
+ fss->n_fl_chunk_bytes += n_collect;
+ fsh_cached_bytes_add (fsh, n_collect);
+
+ clib_spinlock_unlock (&fss->chunk_lock);
+}
+
+void
+fsh_collect_chunks (fifo_segment_header_t * fsh, u32 slice_index,
+ svm_fifo_chunk_t * c)
+{
+ fifo_segment_slice_t *fss;
+ fss = fsh_slice_get (fsh, slice_index);
+ fsh_slice_collect_chunks (fsh, fss, c);
+}
+
+static inline void
+fss_fifo_add_active_list (fifo_segment_slice_t * fss, svm_fifo_t * f)
+{
+ if (fss->fifos)
+ {
+ fss->fifos->prev = f;
+ f->next = fss->fifos;
+ }
+ fss->fifos = f;
+}
+
+static inline void
+fss_fifo_del_active_list (fifo_segment_slice_t * fss, svm_fifo_t * f)
+{
+ if (f->flags & SVM_FIFO_F_LL_TRACKED)
+ {
+ if (f->prev)
+ f->prev->next = f->next;
+ else
+ fss->fifos = f->next;
+ if (f->next)
+ f->next->prev = f->prev;
+ }
+}
+