vlib_buffer_main_t buffer_main;
+/* logging */
+static vlib_log_class_t buffer_log_default;
+
uword
vlib_buffer_length_in_chain_slow_path (vlib_main_t * vm,
vlib_buffer_t * b_first)
vlib_buffer_alloc_from_free_list (vm, &l->next_buffer, 1,
free_list_index))
return copied;
- *last = l = vlib_buffer_chain_buffer (vm, first, l, l->next_buffer);
+ *last = l = vlib_buffer_chain_buffer (vm, l, l->next_buffer);
max = n_buffer_bytes - l->current_length - l->current_data;
}
vlib_buffer_pool_t *p;
vlib_physmem_map_t *m = vlib_physmem_get_map (vm, physmem_map_index);
uword start = pointer_to_uword (m->base);
- uword size = m->n_pages << m->log2_page_size;
+ uword size = (uword) m->n_pages << m->log2_page_size;
if (bm->buffer_mem_size == 0)
{
clib_error_t *error;
u32 physmem_map_index;
u8 pool_index;
+ int log2_page_size = 0;
+
+ buffer_log_default = vlib_log_register_class ("buffer", 0);
if (vlib_buffer_callbacks)
{
&vlib_buffer_delete_free_list_internal;
clib_spinlock_init (&bm->buffer_known_hash_lockp);
- if ((error = vlib_physmem_shared_map_create (vm, "buffers",
- vlib_buffer_physmem_sz, 1,
- &physmem_map_index)))
+retry:
+ error = vlib_physmem_shared_map_create (vm, "buffers",
+ vlib_buffer_physmem_sz,
+ log2_page_size,
+ CLIB_PMALLOC_NUMA_LOCAL,
+ &physmem_map_index);
+
+ if (error && log2_page_size == 0)
+ {
+ vlib_log_warn (buffer_log_default, "%U", format_clib_error, error);
+ clib_error_free (error);
+ vlib_log_warn (buffer_log_default, "falling back to non-hugepage "
+ "backed buffer pool");
+ log2_page_size = min_log2 (clib_mem_get_page_size ());
+ goto retry;
+ }
+
+ if (error)
return error;
pool_index = vlib_buffer_register_physmem_map (vm, physmem_map_index);