{
avf_rxq_t *q = vec_elt_at_index (ad->rxqs, i);
rxq->ring_len = q->size;
- rxq->databuffer_size = VLIB_BUFFER_DATA_SIZE;
+ rxq->databuffer_size = vlib_bufer_get_default_size (vm);
rxq->dma_ring_addr = avf_dma_addr (vm, ad, (void *) q->descs);
avf_reg_write (ad, AVF_QRX_TAIL (i), q->size - 1);
}
/* Call the mempool priv initializer */
priv.mbuf_data_room_size = VLIB_BUFFER_PRE_DATA_SIZE +
- VLIB_BUFFER_DATA_SIZE;
+ vlib_bufer_get_default_size (vm);
priv.mbuf_priv_size = VLIB_BUFFER_HDR_SIZE;
rte_pktmbuf_pool_init (mp, &priv);
rte_pktmbuf_pool_init (nmp, &priv);
/* FIXME bpool bit select per pp */
s = format (s, "pool-%d:%d%c", pp2_id, pp2_id + 8, 0);
bpool_params.match = (char *) s;
- bpool_params.buff_len = VLIB_BUFFER_DATA_SIZE;
+ bpool_params.buff_len = vlib_bufer_get_default_size (vm);
/* FIXME +64 ? */
if (pp2_bpool_init (&bpool_params, &ppif->inqs[0].bpool))
{
memif_main_t *mm = &memif_main;
memif_ring_t *ring;
memif_queue_t *mq;
- u16 buffer_size = VLIB_BUFFER_DATA_SIZE;
+ u16 buffer_size = vlib_bufer_get_default_size (vm);
uword n_trace = vlib_get_trace_count (vm, node);
u16 nexts[MEMIF_RX_VECTOR_SZ], *next = nexts;
u32 _to_next_bufs[MEMIF_RX_VECTOR_SZ], *to_next_bufs = _to_next_bufs, *bi;
/* asume that somebody will want to add ethernet header on the packet
so start with IP header at offset 14 */
start_offset = (mode == MEMIF_INTERFACE_MODE_IP) ? 14 : 0;
- buffer_length = VLIB_BUFFER_DATA_SIZE - start_offset;
+ buffer_length = vlib_bufer_get_default_size (vm) - start_offset;
cur_slot = mq->last_tail;
last_slot = ring->tail;
#define MTU 1500
-#define MTU_BUFFERS ((MTU + VLIB_BUFFER_DATA_SIZE - 1) / VLIB_BUFFER_DATA_SIZE)
+#define MTU_BUFFERS ((MTU + vlib_bufer_get_default_size(vm) - 1) / vlib_bufer_get_default_size(vm))
#define NUM_BUFFERS_TO_ALLOC 32
/*
vlib_buffer_t *b = vlib_get_buffer (vm, ring->bufs[ring->produce]);
rxd = &rxq->rx_desc[0][ring->produce];
rxd->address = vlib_buffer_get_pa (vm, b);
- rxd->flags = ring->gen | VLIB_BUFFER_DATA_SIZE;
+ rxd->flags = ring->gen | vlib_bufer_get_default_size (vm);
vmxnet3_rx_ring_advance_produce (rxq, ring);
ring->fill++;
vlib_buffer_t *b = vlib_get_buffer (vm, ring->bufs[ring->produce]);
rxd = &rxq->rx_desc[1][ring->produce];
rxd->address = vlib_buffer_get_pa (vm, b);
- rxd->flags = ring->gen | VLIB_BUFFER_DATA_SIZE | VMXNET3_RXF_BTYPE;
+ rxd->flags =
+ ring->gen | vlib_bufer_get_default_size (vm) | VMXNET3_RXF_BTYPE;
vmxnet3_rx_ring_advance_produce (rxq, ring);
ring->fill++;
if ((signed) b->current_data < (signed) -VLIB_BUFFER_PRE_DATA_SIZE)
return format (0, "current data %d before pre-data", b->current_data);
- if (b->current_data + b->current_length > VLIB_BUFFER_DATA_SIZE)
+ if (b->current_data + b->current_length > vlib_bufer_get_default_size (vm))
return format (0, "%d-%d beyond end of buffer %d", b->current_data,
- b->current_length, VLIB_BUFFER_DATA_SIZE);
+ b->current_length, vlib_bufer_get_default_size (vm));
if (follow_buffer_next && (b->flags & VLIB_BUFFER_NEXT_PRESENT))
{
d = data;
n_left = n_data_bytes;
- n_buffer_bytes = VLIB_BUFFER_DATA_SIZE;
+ n_buffer_bytes = vlib_bufer_get_default_size (vm);
b = vlib_get_buffer (vm, bi);
b->flags &= ~VLIB_BUFFER_TOTAL_LENGTH_VALID;
u16 data_len)
{
vlib_buffer_t *l = *last;
- u32 n_buffer_bytes = VLIB_BUFFER_DATA_SIZE;
+ u32 n_buffer_bytes = vlib_bufer_get_default_size (vm);
u16 copied = 0;
ASSERT (n_buffer_bytes >= l->current_length + l->current_data);
while (data_len)
u32 buffers_per_numa;
u32 buffer_size = CLIB_CACHE_LINE_ROUND (bm->ext_hdr_size +
sizeof (vlib_buffer_t) +
- VLIB_BUFFER_DATA_SIZE);
+ vlib_bufer_get_default_size (vm));
u8 *name;
pagesize = clib_mem_get_default_hugepage_size ();
name = format (name, "default-numa-%d%c", numa_node, 0);
return vlib_buffer_pool_create (vm, numa_node, (char *) name,
- VLIB_BUFFER_DATA_SIZE, physmem_map_index);
+ vlib_bufer_get_default_size (vm),
+ physmem_map_index);
+}
+
+void
+vlib_buffer_main_alloc (vlib_main_t * vm)
+{
+ vlib_buffer_main_t *bm;
+
+ if (vm->buffer_main)
+ return;
+
+ vm->buffer_main = bm = clib_mem_alloc (sizeof (bm[0]));
+ clib_memset (vm->buffer_main, 0, sizeof (bm[0]));
+ bm->default_data_size = VLIB_BUFFER_DEFAULT_DATA_SIZE;
}
clib_error_t *
clib_bitmap_t *bmp = 0;
u32 numa_node;
- if (vm->buffer_main == 0)
- {
- vm->buffer_main = clib_mem_alloc (sizeof (bm[0]));
- clib_memset (vm->buffer_main, 0, sizeof (bm[0]));
- }
+ vlib_buffer_main_alloc (vm);
bm = vm->buffer_main;
bm->log_default = vlib_log_register_class ("buffer", 0);
{
vlib_buffer_main_t *bm;
- if (vm->buffer_main == 0)
- {
- vm->buffer_main = clib_mem_alloc (sizeof (bm[0]));
- clib_memset (vm->buffer_main, 0, sizeof (bm[0]));
- }
+ vlib_buffer_main_alloc (vm);
bm = vm->buffer_main;
{
if (unformat (input, "buffers-per-numa %u", &bm->buffers_per_numa))
;
+ else if (unformat (input, "default data-size %u",
+ &bm->default_data_size))
+ ;
else
return unformat_parse_error (input);
}
#include <vlib/error.h> /* for vlib_error_t */
#include <vlib/config.h> /* for __PRE_DATA_SIZE */
-#define VLIB_BUFFER_DATA_SIZE (2048)
#define VLIB_BUFFER_PRE_DATA_SIZE __PRE_DATA_SIZE
+#define VLIB_BUFFER_DEFAULT_DATA_SIZE (2048)
+
/* Minimum buffer chain segment size. Does not apply to last buffer in chain.
Dataplane code can safely asume that specified amount of data is not split
into 2 chained buffers */
/* config */
u32 buffers_per_numa;
u16 ext_hdr_size;
+ u32 default_data_size;
/* logging */
vlib_log_class_t log_default;
/* reference count in allocated buffer always must be 1 or higher */
ASSERT (b->ref_count > 0);
- /* verify that buffer pointer is from buffer memory range */
- ASSERT (pointer_to_uword (b) >= bm->buffer_mem_start);
- ASSERT (pointer_to_uword (b) < bm->buffer_mem_start + bm->buffer_mem_size -
- VLIB_BUFFER_DATA_SIZE);
-
/* verify that buffer pool index is valid */
bp = vec_elt_at_index (bm->buffer_pools, b->buffer_pool_index);
ASSERT (pointer_to_uword (b) >= bp->start);
ASSERT (pointer_to_uword (b) < bp->start + bp->size -
- VLIB_BUFFER_DATA_SIZE);
+ (bp->data_size + sizeof (vlib_buffer_t)));
}
always_inline void *
return b;
}
+static_always_inline u32
+vlib_bufer_get_default_size (vlib_main_t * vm)
+{
+ return vm->buffer_main->default_data_size;
+}
+
static_always_inline void
vlib_buffer_copy_indices (u32 * dst, u32 * src, u32 n_indices)
{
vlib_buffer_t * first,
vlib_buffer_t * last, void *data, u16 data_len)
{
- u32 n_buffer_bytes = VLIB_BUFFER_DATA_SIZE;
+ u32 n_buffer_bytes = vlib_bufer_get_default_size (vm);
ASSERT (n_buffer_bytes >= last->current_length + last->current_data);
u16 len = clib_min (data_len,
n_buffer_bytes - last->current_length -
}
u32 want_first_size = clib_min (VLIB_BUFFER_CLONE_HEAD_SIZE,
- VLIB_BUFFER_DATA_SIZE -
+ vlib_bufer_get_default_size (vm) -
first->current_data);
do
{
vlib_buffer_chain_linearize (vlib_main_t * vm, vlib_buffer_t * first)
{
vlib_buffer_t *b = first;
- u32 buf_len = VLIB_BUFFER_DATA_SIZE;
+ u32 buf_len = vlib_bufer_get_default_size (vm);
// free buffer chain starting from the second buffer
int free_count = (b->flags & VLIB_BUFFER_NEXT_PRESENT) != 0;
u32 chain_to_free = b->next_buffer;
vlib_main_t *vm = msm->mc_main.vlib_main;
vlib_buffer_t *b;
uword n_left, n_alloc, n_mtu, i, i_rx;
- const uword buffer_size = VLIB_BUFFER_DATA_SIZE;
+ const uword buffer_size = vlib_bufer_get_default_size (vm);
word n_bytes_left;
/* Make sure we have at least a MTU worth of buffers. */
}
msm->rx_mtu_n_bytes = mtu;
- msm->rx_mtu_n_buffers = msm->rx_mtu_n_bytes / VLIB_BUFFER_DATA_SIZE;
- msm->rx_mtu_n_buffers += (msm->rx_mtu_n_bytes % VLIB_BUFFER_DATA_SIZE) != 0;
+ msm->rx_mtu_n_buffers =
+ msm->rx_mtu_n_bytes / vlib_bufer_get_default_size (vm);
+ msm->rx_mtu_n_buffers +=
+ (msm->rx_mtu_n_bytes % vlib_bufer_get_default_size (vm)) != 0;
error = socket_setup (msm);
if (error)
u8 *block_start = apif->rx_ring + block * block_size;
uword n_trace = vlib_get_trace_count (vm, node);
u32 thread_index = vm->thread_index;
- u32 n_buffer_bytes = VLIB_BUFFER_DATA_SIZE;
+ u32 n_buffer_bytes = vlib_bufer_get_default_size (vm);
u32 min_bufs = apif->rx_req->tp_frame_size / n_buffer_bytes;
if (apif->per_interface_next_index != ~0)
struct netmap_ring *ring;
int cur_ring;
u32 thread_index = vm->thread_index;
- u32 n_buffer_bytes = VLIB_BUFFER_DATA_SIZE;
+ u32 n_buffer_bytes = vlib_bufer_get_default_size (vm);
if (nif->per_interface_next_index != ~0)
next_index = nif->per_interface_next_index;
((vif->type == VIRTIO_IF_TYPE_PCI) ? vlib_buffer_get_current_pa (vm,
b) :
pointer_to_uword (vlib_buffer_get_current (b)));
- d->len = VLIB_BUFFER_DATA_SIZE + hdr_sz;
+ d->len = vlib_bufer_get_default_size (vm) + hdr_sz;
d->flags = VRING_DESC_F_WRITE;
vring->avail->ring[avail & mask] = next;
avail++;
/* Get more output if necessary. Or end of packet. */
if (PREDICT_FALSE
- (b_current->current_length == VLIB_BUFFER_DATA_SIZE))
+ (b_current->current_length == vlib_bufer_get_default_size (vm)))
{
if (PREDICT_FALSE (cpu->rx_buffers_len == 0))
{
vhost_copy_t *cpy = &cpu->copy[copy_len];
copy_len++;
u32 desc_data_l = desc_table[desc_current].len - desc_data_offset;
- cpy->len = VLIB_BUFFER_DATA_SIZE - b_current->current_length;
+ cpy->len =
+ vlib_bufer_get_default_size (vm) - b_current->current_length;
cpy->len = (cpy->len > desc_data_l) ? desc_data_l : cpy->len;
cpy->dst = (uword) (vlib_buffer_get_current (b_current) +
b_current->current_length);
link_address_set:
if ((b0->current_length + sizeof (*id1) + sizeof (*vss1) +
- sizeof (*cmac)) > VLIB_BUFFER_DATA_SIZE)
+ sizeof (*cmac)) > vlib_bufer_get_default_size (vm))
{
error0 = DHCPV6_PROXY_ERROR_PKT_TOO_BIG;
next0 = DHCPV6_PROXY_TO_SERVER_INPUT_NEXT_DROP;
rem = clib_net_to_host_u16 (ip4->length) - sizeof (ip4_header_t);
max =
- (clib_min (mtu, VLIB_BUFFER_DATA_SIZE) - sizeof (ip4_header_t)) & ~0x7;
+ (clib_min (mtu, vlib_bufer_get_default_size (vm)) -
+ sizeof (ip4_header_t)) & ~0x7;
if (rem >
(vlib_buffer_length_in_chain (vm, from_b) - sizeof (ip4_header_t)))
int l34_len =
l4_header_offset + sizeof (icmp46_header_t) +
offsetof (icmp46_echo_request_t, data);
- int max_data_len = VLIB_BUFFER_DATA_SIZE - l34_len;
+ int max_data_len = vlib_bufer_get_default_size (vm) - l34_len;
int first_buf_data_len = data_len < max_data_len ? data_len : max_data_len;
{
int this_buf_data_len =
remaining_data_len <
- VLIB_BUFFER_DATA_SIZE ? remaining_data_len : VLIB_BUFFER_DATA_SIZE;
+ vlib_bufer_get_default_size (vm) ? remaining_data_len :
+ vlib_bufer_get_default_size (vm);
int n_alloc = vlib_buffer_alloc (vm, &b0->next_buffer, 1);
if (n_alloc < 1)
{
static uword
punt_socket_rx_fd (vlib_main_t * vm, vlib_node_runtime_t * node, u32 fd)
{
- const uword buffer_size = VLIB_BUFFER_DATA_SIZE;
+ const uword buffer_size = vlib_bufer_get_default_size (vm);
u32 n_trace = vlib_get_trace_count (vm, node);
u32 next = node->cached_next_index;
u32 n_left_to_next, next_index;
/* Initialize the buffer */
VLIB_BUFFER_TRACE_TRAJECTORY_INIT (b0);
- ASSERT (vec_len (fr->rewrite) < VLIB_BUFFER_DATA_SIZE);
+ ASSERT (vec_len (fr->rewrite) < vlib_bufer_get_default_size (vm));
clib_memcpy_fast (b0->data, fr->rewrite, vec_len (fr->rewrite));
b0->current_data = 0;
s.sw_if_index[VLIB_RX] = s.sw_if_index[VLIB_TX] = ~0;
s.node_index = ~0;
s.max_packet_bytes = s.min_packet_bytes = 64;
- s.buffer_bytes = VLIB_BUFFER_DATA_SIZE;
+ s.buffer_bytes = vlib_bufer_get_default_size (vm);
s.if_id = 0;
pcap_file_name = 0;
while (unformat_check_input (input) != UNFORMAT_END_OF_INPUT)
u32 *buffers;
vlib_main_t *vm = vlib_get_main ();
vnet_main_t *vnm = vnet_get_main ();
+ u32 buf_sz = vlib_bufer_get_default_size (vm);
vnet_interface_main_t *im = &vnm->interface_main;
vnet_sw_interface_t *si;
u8 *d0;
d0 = vec_elt (s->replay_packet_templates, i);
- buffer_alloc_request += (vec_len (d0) + (VLIB_BUFFER_DATA_SIZE - 1))
- / VLIB_BUFFER_DATA_SIZE;
+ buffer_alloc_request += (vec_len (d0) + (buf_sz - 1)) / buf_sz;
i = ((i + 1) == l) ? 0 : i + 1;
n_left--;
/* Copy the data */
while (bytes_to_copy)
{
- bytes_this_chunk = clib_min (bytes_to_copy, VLIB_BUFFER_DATA_SIZE);
+ bytes_this_chunk = clib_min (bytes_to_copy, buf_sz);
ASSERT (current_buffer_index < vec_len (buffers));
b = vlib_get_buffer (vm, buffers[current_buffer_index]);
clib_memcpy_fast (b->data, d0 + data_offset, bytes_this_chunk);
{
int n;
- s->buffer_bytes = VLIB_BUFFER_DATA_SIZE;
+ s->buffer_bytes = vlib_bufer_get_default_size (vm);
n = s->max_packet_bytes / s->buffer_bytes;
n += (s->max_packet_bytes % s->buffer_bytes) != 0;
vec_validate (tm->ip_lookup_tx_frames[0], num_threads - 1);
vec_validate (tm->ip_lookup_tx_frames[1], num_threads - 1);
- tm->bytes_per_buffer = VLIB_BUFFER_DATA_SIZE;
+ tm->bytes_per_buffer = vlib_bufer_get_default_size (vm);
vec_validate (tm->time_now, num_threads - 1);
return error;
ctx->max_len_to_snd = max_segs * ctx->snd_mss;
}
- n_bytes_per_buf = VLIB_BUFFER_DATA_SIZE;
+ n_bytes_per_buf = vlib_bufer_get_default_size (vm);
ASSERT (n_bytes_per_buf > TRANSPORT_MAX_HDRS_LEN);
n_bytes_per_seg = TRANSPORT_MAX_HDRS_LEN + ctx->snd_mss;
ctx->n_bufs_per_seg = ceil ((f64) n_bytes_per_seg / n_bytes_per_buf);
tcp_initialize_timer_wheels (tm);
tcp_initialize_iss_seed (tm);
- tm->bytes_per_buffer = VLIB_BUFFER_DATA_SIZE;
+ tm->bytes_per_buffer = vlib_bufer_get_default_size (vm);
return error;
}
tuntap_main_t *tm = &tuntap_main;
vlib_buffer_t *b;
u32 bi;
- const uword buffer_size = VLIB_BUFFER_DATA_SIZE;
+ const uword buffer_size = vlib_bufer_get_default_size (vm);
u16 thread_index = vm->thread_index;
/** Make sure we have some RX buffers. */
u8 *name;
int flags = IFF_TUN | IFF_NO_PI;
int is_enabled = 0, is_ether = 0, have_normal_interface = 0;
- const uword buffer_size = VLIB_BUFFER_DATA_SIZE;
+ const uword buffer_size = vlib_bufer_get_default_size (vm);
while (unformat_check_input (input) != UNFORMAT_END_OF_INPUT)
{
## large number of interfaces and worker threads. Value is per numa node.
## Default is 16384 (8192 if running unpriviledged)
# buffers-per-numa 128000
+
+ ## Size of buffer data area
+ ## Default is 2048
+ # default data-size 2048
# }
# dpdk {