static_always_inline void
vmxnet3_handle_offload (vmxnet3_rx_comp * rx_comp, vlib_buffer_t * hb,
- u16 * next, u16 gso_size)
+ u16 gso_size)
{
u8 l4_hdr_sz = 0;
- if (gso_size)
- {
- if (rx_comp->flags & VMXNET3_RXCF_TCP)
- {
- tcp_header_t *tcp =
- (tcp_header_t *) (hb->data + vnet_buffer (hb)->l4_hdr_offset);
- l4_hdr_sz = tcp_header_bytes (tcp);
- }
- else if (rx_comp->flags & VMXNET3_RXCF_UDP)
- {
- udp_header_t *udp =
- (udp_header_t *) (hb->data + vnet_buffer (hb)->l4_hdr_offset);
- l4_hdr_sz = sizeof (*udp);
- }
- }
-
if (rx_comp->flags & VMXNET3_RXCF_IP4)
{
ip4_header_t *ip4 = (ip4_header_t *) (hb->data +
hb->flags |= VNET_BUFFER_F_L2_HDR_OFFSET_VALID |
VNET_BUFFER_F_L3_HDR_OFFSET_VALID |
VNET_BUFFER_F_L4_HDR_OFFSET_VALID | VNET_BUFFER_F_IS_IP4;
- next[0] = VNET_DEVICE_INPUT_NEXT_IP4_NCS_INPUT;
/* checksum offload */
if (!(rx_comp->index & VMXNET3_RXCI_CNC))
if (gso_size)
{
+ if (rx_comp->flags & VMXNET3_RXCF_TCP)
+ {
+ tcp_header_t *tcp =
+ (tcp_header_t *) (hb->data + vnet_buffer (hb)->l4_hdr_offset);
+ l4_hdr_sz = tcp_header_bytes (tcp);
+ }
+ else if (rx_comp->flags & VMXNET3_RXCF_UDP)
+ {
+ udp_header_t *udp =
+ (udp_header_t *) (hb->data + vnet_buffer (hb)->l4_hdr_offset);
+ l4_hdr_sz = sizeof (*udp);
+ }
vnet_buffer2 (hb)->gso_size = gso_size;
vnet_buffer2 (hb)->gso_l4_hdr_sz = l4_hdr_sz;
hb->flags |= VNET_BUFFER_F_GSO;
}
- vlib_buffer_advance (hb, device_input_next_node_advance[next[0]]);
}
else if (rx_comp->flags & VMXNET3_RXCF_IP6)
{
hb->flags |= VNET_BUFFER_F_L2_HDR_OFFSET_VALID |
VNET_BUFFER_F_L3_HDR_OFFSET_VALID |
VNET_BUFFER_F_L4_HDR_OFFSET_VALID | VNET_BUFFER_F_IS_IP6;
- next[0] = VNET_DEVICE_INPUT_NEXT_IP6_INPUT;
/* checksum offload */
if (!(rx_comp->index & VMXNET3_RXCI_CNC))
if (gso_size)
{
+ if (rx_comp->flags & VMXNET3_RXCF_TCP)
+ {
+ tcp_header_t *tcp =
+ (tcp_header_t *) (hb->data + vnet_buffer (hb)->l4_hdr_offset);
+ l4_hdr_sz = tcp_header_bytes (tcp);
+ }
+ else if (rx_comp->flags & VMXNET3_RXCF_UDP)
+ {
+ udp_header_t *udp =
+ (udp_header_t *) (hb->data + vnet_buffer (hb)->l4_hdr_offset);
+ l4_hdr_sz = sizeof (*udp);
+ }
vnet_buffer2 (hb)->gso_size = gso_size;
vnet_buffer2 (hb)->gso_l4_hdr_sz = l4_hdr_sz;
hb->flags |= VNET_BUFFER_F_GSO;
}
- vlib_buffer_advance (hb, device_input_next_node_advance[next[0]]);
}
- else
- next[0] = VNET_DEVICE_INPUT_NEXT_ETHERNET_INPUT;
}
static_always_inline uword
b0->flags = 0;
b0->error = 0;
b0->current_config_index = 0;
- ASSERT (b0->current_length != 0);
if (PREDICT_FALSE ((rx_comp->index & VMXNET3_RXCI_EOP) &&
(rx_comp->len & VMXNET3_RXCL_ERROR)))
{
ASSERT (!(rxd->flags & VMXNET3_RXF_BTYPE));
/* start segment */
- if ((vd->lro_enable) &&
+ if (vd->gso_enable &&
(rx_comp->flags & VMXNET3_RXCF_CT) == VMXNET3_RXCOMP_TYPE_LRO)
{
vmxnet3_rx_comp_ext *lro = (vmxnet3_rx_comp_ext *) rx_comp;
else if (rx_comp->index & VMXNET3_RXCI_EOP)
{
/* end of segment */
- if (prev_b0)
+ if (PREDICT_TRUE (prev_b0 != 0))
{
- prev_b0->flags |= VLIB_BUFFER_NEXT_PRESENT;
- prev_b0->next_buffer = bi0;
- hb->total_length_not_including_first_buffer +=
- b0->current_length;
+ if (PREDICT_TRUE (b0->current_length != 0))
+ {
+ prev_b0->flags |= VLIB_BUFFER_NEXT_PRESENT;
+ prev_b0->next_buffer = bi0;
+ hb->total_length_not_including_first_buffer +=
+ b0->current_length;
+ }
+ else
+ {
+ vlib_buffer_free_one (vm, bi0);
+ }
prev_b0 = 0;
got_packet = 1;
}
}
if (PREDICT_FALSE (known_next))
- {
- next[0] = next_index;
- }
+ next[0] = next_index;
else
{
ethernet_header_t *e = (ethernet_header_t *) hb->data;
- if (ethernet_frame_is_tagged (e->type))
- next[0] = VNET_DEVICE_INPUT_NEXT_ETHERNET_INPUT;
- else
- vmxnet3_handle_offload (rx_comp, hb, next, gso_size);
+ next[0] = VNET_DEVICE_INPUT_NEXT_ETHERNET_INPUT;
+ if (!ethernet_frame_is_tagged (ntohs (e->type)))
+ vmxnet3_handle_offload (rx_comp, hb, gso_size);
}
n_rx_packets++;
next = nexts;
while (n_trace && n_left)
{
- vlib_buffer_t *b;
- vmxnet3_input_trace_t *tr;
-
- b = vlib_get_buffer (vm, bi[0]);
- vlib_trace_buffer (vm, node, next[0], b, /* follow_chain */ 0);
- tr = vlib_add_trace (vm, node, b, sizeof (*tr));
- tr->next_index = next[0];
- tr->hw_if_index = vd->hw_if_index;
- tr->buffer = *b;
-
- n_trace--;
+ vlib_buffer_t *b = vlib_get_buffer (vm, bi[0]);
+ if (PREDICT_TRUE
+ (vlib_trace_buffer
+ (vm, node, next[0], b, /* follow_chain */ 0)))
+ {
+ vmxnet3_input_trace_t *tr =
+ vlib_add_trace (vm, node, b, sizeof (*tr));
+ tr->next_index = next[0];
+ tr->hw_if_index = vd->hw_if_index;
+ tr->buffer = *b;
+ n_trace--;
+ }
n_left--;
bi++;
next++;
vlib_increment_combined_counter
(vnm->interface_main.combined_sw_if_counters +
VNET_INTERFACE_COUNTER_RX, thread_index,
- vd->hw_if_index, n_rx_packets, n_rx_bytes);
+ vd->sw_if_index, n_rx_packets, n_rx_bytes);
}
error = vmxnet3_rxq_refill_ring0 (vm, vd, rxq);
VLIB_REGISTER_NODE (vmxnet3_input_node) = {
.name = "vmxnet3-input",
.sibling_of = "device-input",
+ .flags = VLIB_NODE_FLAG_TRACE_SUPPORTED,
.format_trace = format_vmxnet3_input_trace,
.type = VLIB_NODE_TYPE_INPUT,
.state = VLIB_NODE_STATE_DISABLED,