#include <vlib/vlib.h>
#include <vlib/unix/unix.h>
#include <vnet/ethernet/ethernet.h>
-#include <vnet/devices/devices.h>
+#include <vnet/interface/rx_queue_funcs.h>
#include <vnet/feature/feature.h>
#include <memif/memif.h>
#include <memif/private.h>
-#define foreach_memif_input_error \
- _(BUFFER_ALLOC_FAIL, "buffer allocation failed") \
- _(NOT_IP, "not ip packet")
+#define foreach_memif_input_error \
+ _ (BUFFER_ALLOC_FAIL, "buffer allocation failed") \
+ _ (BAD_DESC, "bad descriptor") \
+ _ (NOT_IP, "not ip packet")
typedef enum
{
{
VLIB_BUFFER_TRACE_TRAJECTORY_INIT (b);
- if (PREDICT_TRUE (b != 0))
+ if (PREDICT_TRUE
+ (b != 0 && vlib_trace_buffer (vm, node, next, b, /* follow_chain */ 0)))
{
memif_input_trace_t *tr;
- vlib_trace_buffer (vm, node, next, b, /* follow_chain */ 0);
vlib_set_trace_count (vm, node, --(*n_tracep));
tr = vlib_add_trace (vm, node, b, sizeof (*tr));
tr->next_index = next;
memif_main_t *mm = &memif_main;
memif_ring_t *ring;
memif_queue_t *mq;
- u16 buffer_size = VLIB_BUFFER_DATA_SIZE;
+ u16 buffer_size = vlib_buffer_get_default_data_size (vm);
uword n_trace = vlib_get_trace_count (vm, node);
u16 nexts[MEMIF_RX_VECTOR_SZ], *next = nexts;
u32 _to_next_bufs[MEMIF_RX_VECTOR_SZ], *to_next_bufs = _to_next_bufs, *bi;
u32 n_rx_packets = 0, n_rx_bytes = 0;
- u32 n_left, n_left_to_next, next_index;
+ u32 n_left, n_left_to_next;
+ u32 next_index = VNET_DEVICE_INPUT_NEXT_ETHERNET_INPUT;
vlib_buffer_t *b0, *b1, *b2, *b3;
u32 thread_index = vm->thread_index;
memif_per_thread_data_t *ptd = vec_elt_at_index (mm->per_thread_data,
memif_packet_op_t *po;
memif_region_index_t last_region = ~0;
void *last_region_shm = 0;
+ void *last_region_max = 0;
mq = vec_elt_at_index (mif->rx_queues, qid);
ring = mq->ring;
ring_size = 1 << mq->log2_ring_size;
mask = ring_size - 1;
- /* asume that somebody will want to add ethernet header on the packet
+ /* assume that somebody will want to add ethernet header on the packet
so start with IP header at offset 14 */
start_offset = (mode == MEMIF_INTERFACE_MODE_IP) ? 14 : 0;
{
last_region_shm = mif->regions[d0->region].shm;
last_region = d0->region;
+ last_region_max =
+ last_region_shm + mif->regions[last_region].region_size;
}
mb0 = last_region_shm + d0->offset;
- do
- {
- u32 dst_free = buffer_size - dst_off;
- if (dst_free == 0)
- {
- dst_off = 0;
- dst_free = buffer_size;
- n_buffers++;
- }
- u32 bytes_to_copy = clib_min (dst_free, n_bytes_left);
- memif_add_copy_op (ptd, mb0 + src_off, bytes_to_copy, dst_off,
- n_buffers - 1);
- n_bytes_left -= bytes_to_copy;
- src_off += bytes_to_copy;
- dst_off += bytes_to_copy;
- }
- while (PREDICT_FALSE (n_bytes_left));
+ if (PREDICT_FALSE (mb0 + n_bytes_left > last_region_max))
+ vlib_error_count (vm, node->node_index, MEMIF_INPUT_ERROR_BAD_DESC, 1);
+ else
+ do
+ {
+ u32 dst_free = buffer_size - dst_off;
+ if (dst_free == 0)
+ {
+ dst_off = 0;
+ dst_free = buffer_size;
+ n_buffers++;
+ }
+ u32 bytes_to_copy = clib_min (dst_free, n_bytes_left);
+ memif_add_copy_op (ptd, mb0 + src_off, bytes_to_copy, dst_off,
+ n_buffers - 1);
+ n_bytes_left -= bytes_to_copy;
+ src_off += bytes_to_copy;
+ dst_off += bytes_to_copy;
+ }
+ while (PREDICT_FALSE (n_bytes_left));
cur_slot++;
n_slots--;
ptd->buffer_template.current_data = start_offset;
ptd->buffer_template.current_config_index = 0;
ptd->buffer_template.buffer_pool_index = mq->buffer_pool_index;
+ ptd->buffer_template.ref_count = 1;
if (mode == MEMIF_INTERFACE_MODE_ETHERNET)
{
vlib_frame_t *f;
ethernet_input_frame_t *ef;
nf = vlib_node_runtime_get_next_frame (vm, node, next_index);
- f = vlib_get_frame (vm, nf->frame_index);
+ f = vlib_get_frame (vm, nf->frame);
f->flags = ETH_INPUT_FRAME_F_SINGLE_SW_IF_IDX;
ef = vlib_frame_scalar_args (f);
ef->sw_if_index = mif->sw_if_index;
ef->hw_if_index = mif->hw_if_index;
+ vlib_frame_no_append (f);
}
}
while (n_from >= 8)
{
- b0 = vlib_get_buffer (vm, po[4].first_buffer_vec_index);
- b1 = vlib_get_buffer (vm, po[5].first_buffer_vec_index);
- b2 = vlib_get_buffer (vm, po[6].first_buffer_vec_index);
- b3 = vlib_get_buffer (vm, po[7].first_buffer_vec_index);
+ b0 = vlib_get_buffer (vm, ptd->buffers[po[0].first_buffer_vec_index]);
+ b1 = vlib_get_buffer (vm, ptd->buffers[po[1].first_buffer_vec_index]);
+ b2 = vlib_get_buffer (vm, ptd->buffers[po[2].first_buffer_vec_index]);
+ b3 = vlib_get_buffer (vm, ptd->buffers[po[3].first_buffer_vec_index]);
+
vlib_prefetch_buffer_header (b0, STORE);
vlib_prefetch_buffer_header (b1, STORE);
vlib_prefetch_buffer_header (b2, STORE);
if (mode != MEMIF_INTERFACE_MODE_ETHERNET)
ni = next[0];
b = vlib_get_buffer (vm, bi[0]);
- vlib_trace_buffer (vm, node, ni, b, /* follow_chain */ 0);
- tr = vlib_add_trace (vm, node, b, sizeof (*tr));
- tr->next_index = ni;
- tr->hw_if_index = mif->hw_if_index;
- tr->ring = qid;
+ if (PREDICT_TRUE
+ (vlib_trace_buffer (vm, node, ni, b, /* follow_chain */ 0)))
+ {
+ tr = vlib_add_trace (vm, node, b, sizeof (*tr));
+ tr->next_index = ni;
+ tr->hw_if_index = mif->hw_if_index;
+ tr->ring = qid;
+ n_trace--;
+ }
/* next */
- n_trace--;
n_left--;
bi++;
next++;
vlib_increment_combined_counter (vnm->interface_main.combined_sw_if_counters
+ VNET_INTERFACE_COUNTER_RX, thread_index,
- mif->hw_if_index, n_rx_packets,
+ mif->sw_if_index, n_rx_packets,
n_rx_bytes);
/* refill ring with empty buffers */
/* asume that somebody will want to add ethernet header on the packet
so start with IP header at offset 14 */
start_offset = (mode == MEMIF_INTERFACE_MODE_IP) ? 14 : 0;
- buffer_length = VLIB_BUFFER_DATA_SIZE - start_offset;
+ buffer_length = vlib_buffer_get_default_data_size (vm) - start_offset;
cur_slot = mq->last_tail;
last_slot = ring->tail;
d0 = &ring->desc[s0];
hb = b0 = vlib_get_buffer (vm, bi0);
b0->current_data = start_offset;
- b0->current_length = start_offset + d0->length;
+ b0->current_length = d0->length;
n_rx_bytes += d0->length;
if (0 && memif_desc_is_invalid (mif, d0, buffer_length))
/* current buffer */
b0 = vlib_get_buffer (vm, bi0);
b0->current_data = start_offset;
- b0->current_length = start_offset + d0->length;
+ b0->current_length = d0->length;
hb->total_length_not_including_first_buffer += d0->length;
n_rx_bytes += d0->length;
vlib_increment_combined_counter (vnm->interface_main.combined_sw_if_counters
+ VNET_INTERFACE_COUNTER_RX, thread_index,
- mif->hw_if_index, n_rx_packets,
+ mif->sw_if_index, n_rx_packets,
n_rx_bytes);
/* refill ring with empty buffers */
{
u32 n_rx = 0;
memif_main_t *mm = &memif_main;
- vnet_device_input_runtime_t *rt = (void *) node->runtime_data;
- vnet_device_and_queue_t *dq;
memif_interface_mode_t mode_ip = MEMIF_INTERFACE_MODE_IP;
memif_interface_mode_t mode_eth = MEMIF_INTERFACE_MODE_ETHERNET;
- foreach_device_and_queue (dq, rt->devices_and_queues)
- {
- memif_if_t *mif;
- mif = vec_elt_at_index (mm->interfaces, dq->dev_instance);
- if ((mif->flags & MEMIF_IF_FLAG_ADMIN_UP) &&
- (mif->flags & MEMIF_IF_FLAG_CONNECTED))
- {
- if (mif->flags & MEMIF_IF_FLAG_ZERO_COPY)
- {
- if (mif->mode == MEMIF_INTERFACE_MODE_IP)
- n_rx += memif_device_input_zc_inline (vm, node, frame, mif,
- dq->queue_id, mode_ip);
- else
- n_rx += memif_device_input_zc_inline (vm, node, frame, mif,
- dq->queue_id, mode_eth);
- }
- else if (mif->flags & MEMIF_IF_FLAG_IS_SLAVE)
- {
- if (mif->mode == MEMIF_INTERFACE_MODE_IP)
- n_rx += memif_device_input_inline (vm, node, frame, mif,
- MEMIF_RING_M2S, dq->queue_id,
- mode_ip);
- else
- n_rx += memif_device_input_inline (vm, node, frame, mif,
- MEMIF_RING_M2S, dq->queue_id,
- mode_eth);
- }
- else
- {
- if (mif->mode == MEMIF_INTERFACE_MODE_IP)
- n_rx += memif_device_input_inline (vm, node, frame, mif,
- MEMIF_RING_S2M, dq->queue_id,
- mode_ip);
- else
- n_rx += memif_device_input_inline (vm, node, frame, mif,
- MEMIF_RING_S2M, dq->queue_id,
- mode_eth);
- }
- }
- }
+ vnet_hw_if_rxq_poll_vector_t *pv;
+ pv = vnet_hw_if_get_rxq_poll_vector (vm, node);
+ for (int i = 0; i < vec_len (pv); i++)
+ {
+ memif_if_t *mif;
+ u32 qid;
+ mif = vec_elt_at_index (mm->interfaces, pv[i].dev_instance);
+ qid = pv[i].queue_id;
+ if ((mif->flags & MEMIF_IF_FLAG_ADMIN_UP) &&
+ (mif->flags & MEMIF_IF_FLAG_CONNECTED))
+ {
+ if (mif->flags & MEMIF_IF_FLAG_ZERO_COPY)
+ {
+ if (mif->mode == MEMIF_INTERFACE_MODE_IP)
+ n_rx += memif_device_input_zc_inline (vm, node, frame, mif,
+ qid, mode_ip);
+ else
+ n_rx += memif_device_input_zc_inline (vm, node, frame, mif,
+ qid, mode_eth);
+ }
+ else if (mif->flags & MEMIF_IF_FLAG_IS_SLAVE)
+ {
+ if (mif->mode == MEMIF_INTERFACE_MODE_IP)
+ n_rx += memif_device_input_inline (
+ vm, node, frame, mif, MEMIF_RING_M2S, qid, mode_ip);
+ else
+ n_rx += memif_device_input_inline (
+ vm, node, frame, mif, MEMIF_RING_M2S, qid, mode_eth);
+ }
+ else
+ {
+ if (mif->mode == MEMIF_INTERFACE_MODE_IP)
+ n_rx += memif_device_input_inline (
+ vm, node, frame, mif, MEMIF_RING_S2M, qid, mode_ip);
+ else
+ n_rx += memif_device_input_inline (
+ vm, node, frame, mif, MEMIF_RING_S2M, qid, mode_eth);
+ }
+ }
+ }
return n_rx;
}
/* *INDENT-OFF* */
VLIB_REGISTER_NODE (memif_input_node) = {
.name = "memif-input",
+ .flags = VLIB_NODE_FLAG_TRACE_SUPPORTED,
.sibling_of = "device-input",
.format_trace = format_memif_input_trace,
.type = VLIB_NODE_TYPE_INPUT,