n_alloc_per_page = (1ULL << m->log2_page_size) / alloc_size;
/* preallocate buffer indices memory */
- vec_validate_aligned (bp->buffers, m->n_pages * n_alloc_per_page,
- CLIB_CACHE_LINE_BYTES);
- vec_reset_length (bp->buffers);
+ bp->n_buffers = m->n_pages * n_alloc_per_page;
+ bp->buffers = clib_mem_alloc_aligned (bp->n_buffers * sizeof (u32),
+ CLIB_CACHE_LINE_BYTES);
clib_spinlock_init (&bp->lock);
bi = vlib_get_buffer_index (vm, (vlib_buffer_t *) p);
- vec_add1_aligned (bp->buffers, bi, CLIB_CACHE_LINE_BYTES);
+ bp->buffers[bp->n_avail++] = bi;
+
vlib_get_buffer (vm, bi);
}
- bp->n_buffers = vec_len (bp->buffers);
return bp->index;
}
/* *INDENT-OFF* */
vec_foreach (bpt, bp->threads)
- cached += vec_len (bpt->cached_buffers);
+ cached += bpt->n_cached;
/* *INDENT-ON* */
s = format (s, "%-20s%=6d%=6d%=6u%=11u%=6u%=8u%=8u%=8u",
bp->name, bp->index, bp->numa_node, bp->data_size +
sizeof (vlib_buffer_t) + vm->buffer_main->ext_hdr_size,
- bp->data_size, bp->n_buffers, vec_len (bp->buffers), cached,
- bp->n_buffers - vec_len (bp->buffers) - cached);
+ bp->data_size, bp->n_buffers, bp->n_avail, cached,
+ bp->n_buffers - bp->n_avail - cached);
return s;
}
/* *INDENT-OFF* */
vec_foreach (bpt, bp->threads)
- cached += vec_len (bpt->cached_buffers);
+ cached += bpt->n_cached;
/* *INDENT-ON* */
clib_spinlock_unlock (&bp->lock);
if (!bp)
return;
- e->value = bp->n_buffers - vec_len (bp->buffers) - buffer_get_cached (bp);
+ e->value = bp->n_buffers - bp->n_avail - buffer_get_cached (bp);
}
static void
if (!bp)
return;
- e->value = vec_len (bp->buffers);
+ e->value = bp->n_avail;
}
static void