X-Git-Url: https://gerrit.fd.io/r/gitweb?a=blobdiff_plain;f=src%2Fplugins%2Fvmxnet3%2Finput.c;h=25632546b6dde17f246fd6939d6a26de63050bc3;hb=e8fb3cfe892ca1c562d63eb8854386180a297d80;hp=9392d57747d891996aba375ab3784ccd68cb9a04;hpb=8b0995366110ff8c97d1d10aaa8291ad465b0b2f;p=vpp.git diff --git a/src/plugins/vmxnet3/input.c b/src/plugins/vmxnet3/input.c index 9392d57747d..25632546b6d 100644 --- a/src/plugins/vmxnet3/input.c +++ b/src/plugins/vmxnet3/input.c @@ -20,7 +20,11 @@ #include #include #include - +#include +#include +#include +#include +#include #include #define foreach_vmxnet3_input_error \ @@ -71,6 +75,123 @@ vmxnet3_rx_comp_ring_advance_next (vmxnet3_rxq_t * rxq) } } +static_always_inline void +vmxnet3_handle_offload (vmxnet3_rx_comp * rx_comp, vlib_buffer_t * hb, + u16 gso_size) +{ + u8 l4_hdr_sz = 0; + vnet_buffer_oflags_t oflags = 0; + + if (rx_comp->flags & VMXNET3_RXCF_IP4) + { + ip4_header_t *ip4 = (ip4_header_t *) (hb->data + + sizeof (ethernet_header_t)); + + vnet_buffer (hb)->l2_hdr_offset = 0; + vnet_buffer (hb)->l3_hdr_offset = sizeof (ethernet_header_t); + vnet_buffer (hb)->l4_hdr_offset = sizeof (ethernet_header_t) + + ip4_header_bytes (ip4); + hb->flags |= VNET_BUFFER_F_L2_HDR_OFFSET_VALID | + VNET_BUFFER_F_L3_HDR_OFFSET_VALID | + VNET_BUFFER_F_L4_HDR_OFFSET_VALID | VNET_BUFFER_F_IS_IP4; + + /* checksum offload */ + if (!(rx_comp->index & VMXNET3_RXCI_CNC)) + { + if (!(rx_comp->flags & VMXNET3_RXCF_IPC)) + { + oflags |= VNET_BUFFER_OFFLOAD_F_IP_CKSUM; + ip4->checksum = 0; + } + if (!(rx_comp->flags & VMXNET3_RXCF_TUC)) + { + if (rx_comp->flags & VMXNET3_RXCF_TCP) + { + oflags |= VNET_BUFFER_OFFLOAD_F_TCP_CKSUM; + } + else if (rx_comp->flags & VMXNET3_RXCF_UDP) + { + oflags |= VNET_BUFFER_OFFLOAD_F_UDP_CKSUM; + } + } + } + + if (gso_size) + { + if (rx_comp->flags & VMXNET3_RXCF_TCP) + { + tcp_header_t *tcp = + (tcp_header_t *) (hb->data + vnet_buffer (hb)->l4_hdr_offset); + l4_hdr_sz = tcp_header_bytes (tcp); + } + else if (rx_comp->flags & VMXNET3_RXCF_UDP) + { + udp_header_t *udp = + (udp_header_t *) (hb->data + vnet_buffer (hb)->l4_hdr_offset); + l4_hdr_sz = sizeof (*udp); + } + vnet_buffer2 (hb)->gso_size = gso_size; + vnet_buffer2 (hb)->gso_l4_hdr_sz = l4_hdr_sz; + hb->flags |= VNET_BUFFER_F_GSO; + } + } + else if (rx_comp->flags & VMXNET3_RXCF_IP6) + { + vnet_buffer (hb)->l2_hdr_offset = 0; + vnet_buffer (hb)->l3_hdr_offset = sizeof (ethernet_header_t); + vnet_buffer (hb)->l4_hdr_offset = sizeof (ethernet_header_t) + + sizeof (ip6_header_t); + hb->flags |= VNET_BUFFER_F_L2_HDR_OFFSET_VALID | + VNET_BUFFER_F_L3_HDR_OFFSET_VALID | + VNET_BUFFER_F_L4_HDR_OFFSET_VALID | VNET_BUFFER_F_IS_IP6; + + /* checksum offload */ + if (!(rx_comp->index & VMXNET3_RXCI_CNC)) + { + if (!(rx_comp->flags & VMXNET3_RXCF_TUC)) + { + if (rx_comp->flags & VMXNET3_RXCF_TCP) + { + tcp_header_t *tcp = + (tcp_header_t *) (hb->data + + vnet_buffer (hb)->l4_hdr_offset); + oflags |= VNET_BUFFER_OFFLOAD_F_TCP_CKSUM; + tcp->checksum = 0; + } + else if (rx_comp->flags & VMXNET3_RXCF_UDP) + { + udp_header_t *udp = + (udp_header_t *) (hb->data + + vnet_buffer (hb)->l4_hdr_offset); + oflags |= VNET_BUFFER_OFFLOAD_F_UDP_CKSUM; + udp->checksum = 0; + } + } + } + + if (gso_size) + { + if (rx_comp->flags & VMXNET3_RXCF_TCP) + { + tcp_header_t *tcp = + (tcp_header_t *) (hb->data + vnet_buffer (hb)->l4_hdr_offset); + l4_hdr_sz = tcp_header_bytes (tcp); + } + else if (rx_comp->flags & VMXNET3_RXCF_UDP) + { + udp_header_t *udp = + (udp_header_t *) (hb->data + vnet_buffer (hb)->l4_hdr_offset); + l4_hdr_sz = sizeof (*udp); + } + vnet_buffer2 (hb)->gso_size = gso_size; + vnet_buffer2 (hb)->gso_l4_hdr_sz = l4_hdr_sz; + hb->flags |= VNET_BUFFER_F_GSO; + } + } + if (oflags) + vnet_buffer_offload_flags_set (hb, oflags); +} + static_always_inline uword vmxnet3_device_input_inline (vlib_main_t * vm, vlib_node_runtime_t * node, vlib_frame_t * frame, vmxnet3_device_t * vd, @@ -93,6 +214,7 @@ vmxnet3_device_input_inline (vlib_main_t * vm, vlib_node_runtime_t * node, u8 known_next = 0, got_packet = 0; vmxnet3_rx_desc *rxd; clib_error_t *error; + u16 gso_size = 0; rxq = vec_elt_at_index (vd->rxqs, qid); comp_ring = &rxq->rx_comp_ring; @@ -100,8 +222,9 @@ vmxnet3_device_input_inline (vlib_main_t * vm, vlib_node_runtime_t * node, next = nexts; rx_comp = &rxq->rx_comp[comp_ring->next]; - while (PREDICT_TRUE (n_rx_packets < VLIB_FRAME_SIZE) && - (comp_ring->gen == (rx_comp->flags & VMXNET3_RXCF_GEN))) + while (PREDICT_TRUE ((n_rx_packets < VLIB_FRAME_SIZE) && + (comp_ring->gen == + (rx_comp->flags & VMXNET3_RXCF_GEN)))) { vlib_buffer_t *b0; u32 bi0; @@ -142,7 +265,6 @@ vmxnet3_device_input_inline (vlib_main_t * vm, vlib_node_runtime_t * node, b0->flags = 0; b0->error = 0; b0->current_config_index = 0; - ASSERT (b0->current_length != 0); if (PREDICT_FALSE ((rx_comp->index & VMXNET3_RXCI_EOP) && (rx_comp->len & VMXNET3_RXCL_ERROR))) @@ -163,6 +285,14 @@ vmxnet3_device_input_inline (vlib_main_t * vm, vlib_node_runtime_t * node, { ASSERT (!(rxd->flags & VMXNET3_RXF_BTYPE)); /* start segment */ + if (vd->gso_enable && + (rx_comp->flags & VMXNET3_RXCF_CT) == VMXNET3_RXCOMP_TYPE_LRO) + { + vmxnet3_rx_comp_ext *lro = (vmxnet3_rx_comp_ext *) rx_comp; + + gso_size = lro->flags & VMXNET3_RXECF_MSS_MASK; + } + hb = b0; bi[0] = bi0; if (!(rx_comp->index & VMXNET3_RXCI_EOP)) @@ -182,12 +312,19 @@ vmxnet3_device_input_inline (vlib_main_t * vm, vlib_node_runtime_t * node, else if (rx_comp->index & VMXNET3_RXCI_EOP) { /* end of segment */ - if (prev_b0) + if (PREDICT_TRUE (prev_b0 != 0)) { - prev_b0->flags |= VLIB_BUFFER_NEXT_PRESENT; - prev_b0->next_buffer = bi0; - hb->total_length_not_including_first_buffer += - b0->current_length; + if (PREDICT_TRUE (b0->current_length != 0)) + { + prev_b0->flags |= VLIB_BUFFER_NEXT_PRESENT; + prev_b0->next_buffer = bi0; + hb->total_length_not_including_first_buffer += + b0->current_length; + } + else + { + vlib_buffer_free_one (vm, bi0); + } prev_b0 = 0; got_packet = 1; } @@ -231,8 +368,6 @@ vmxnet3_device_input_inline (vlib_main_t * vm, vlib_node_runtime_t * node, if (got_packet) { - ethernet_header_t *e = (ethernet_header_t *) hb->data; - if (PREDICT_FALSE (vd->per_interface_next_index != ~0)) { next_index = vd->per_interface_next_index; @@ -242,42 +377,20 @@ vmxnet3_device_input_inline (vlib_main_t * vm, vlib_node_runtime_t * node, if (PREDICT_FALSE (vnet_device_input_have_features (vd->sw_if_index))) { - vnet_feature_start_device_input_x1 (vd->sw_if_index, - &next_index, hb); + vnet_feature_start_device_input (vd->sw_if_index, &next_index, + hb); known_next = 1; } if (PREDICT_FALSE (known_next)) - { - next[0] = next_index; - } + next[0] = next_index; else { - if (ethernet_frame_is_tagged (e->type)) - next[0] = VNET_DEVICE_INPUT_NEXT_ETHERNET_INPUT; - else - { - if (rx_comp->flags & VMXNET3_RXCF_IP4) - { - next[0] = VNET_DEVICE_INPUT_NEXT_IP4_NCS_INPUT; - hb->flags |= VNET_BUFFER_F_IS_IP4; - vlib_buffer_advance (hb, - device_input_next_node_advance - [next[0]]); - } - else if (rx_comp->flags & VMXNET3_RXCF_IP6) - { - next[0] = VNET_DEVICE_INPUT_NEXT_IP6_INPUT; - hb->flags |= VNET_BUFFER_F_IS_IP6; - vlib_buffer_advance (hb, - device_input_next_node_advance - [next[0]]); - } - else - { - next[0] = VNET_DEVICE_INPUT_NEXT_ETHERNET_INPUT; - } - } + ethernet_header_t *e = (ethernet_header_t *) hb->data; + + next[0] = VNET_DEVICE_INPUT_NEXT_ETHERNET_INPUT; + if (!ethernet_frame_is_tagged (ntohs (e->type))) + vmxnet3_handle_offload (rx_comp, hb, gso_size); } n_rx_packets++; @@ -285,6 +398,7 @@ vmxnet3_device_input_inline (vlib_main_t * vm, vlib_node_runtime_t * node, bi++; hb = 0; got_packet = 0; + gso_size = 0; } next: @@ -300,17 +414,18 @@ vmxnet3_device_input_inline (vlib_main_t * vm, vlib_node_runtime_t * node, next = nexts; while (n_trace && n_left) { - vlib_buffer_t *b; - vmxnet3_input_trace_t *tr; - - b = vlib_get_buffer (vm, bi[0]); - vlib_trace_buffer (vm, node, next[0], b, /* follow_chain */ 0); - tr = vlib_add_trace (vm, node, b, sizeof (*tr)); - tr->next_index = next[0]; - tr->hw_if_index = vd->hw_if_index; - tr->buffer = *b; - - n_trace--; + vlib_buffer_t *b = vlib_get_buffer (vm, bi[0]); + if (PREDICT_TRUE + (vlib_trace_buffer + (vm, node, next[0], b, /* follow_chain */ 0))) + { + vmxnet3_input_trace_t *tr = + vlib_add_trace (vm, node, b, sizeof (*tr)); + tr->next_index = next[0]; + tr->hw_if_index = vd->hw_if_index; + tr->buffer = *b; + n_trace--; + } n_left--; bi++; next++; @@ -325,7 +440,7 @@ vmxnet3_device_input_inline (vlib_main_t * vm, vlib_node_runtime_t * node, vlib_increment_combined_counter (vnm->interface_main.combined_sw_if_counters + VNET_INTERFACE_COUNTER_RX, thread_index, - vd->hw_if_index, n_rx_packets, n_rx_bytes); + vd->sw_if_index, n_rx_packets, n_rx_bytes); } error = vmxnet3_rxq_refill_ring0 (vm, vd, rxq); @@ -350,25 +465,25 @@ VLIB_NODE_FN (vmxnet3_input_node) (vlib_main_t * vm, { u32 n_rx = 0; vmxnet3_main_t *vmxm = &vmxnet3_main; - vnet_device_input_runtime_t *rt = (void *) node->runtime_data; - vnet_device_and_queue_t *dq; - - foreach_device_and_queue (dq, rt->devices_and_queues) - { - vmxnet3_device_t *vd; - vd = vec_elt_at_index (vmxm->devices, dq->dev_instance); - if ((vd->flags & VMXNET3_DEVICE_F_ADMIN_UP) == 0) - continue; - n_rx += vmxnet3_device_input_inline (vm, node, frame, vd, dq->queue_id); - } + vnet_hw_if_rxq_poll_vector_t *pv = vnet_hw_if_get_rxq_poll_vector (vm, node); + vnet_hw_if_rxq_poll_vector_t *pve; + + vec_foreach (pve, pv) + { + vmxnet3_device_t *vd; + vd = vec_elt_at_index (vmxm->devices, pve->dev_instance); + if ((vd->flags & VMXNET3_DEVICE_F_ADMIN_UP) == 0) + continue; + n_rx += vmxnet3_device_input_inline (vm, node, frame, vd, pve->queue_id); + } return n_rx; } #ifndef CLIB_MARCH_VARIANT -/* *INDENT-OFF* */ VLIB_REGISTER_NODE (vmxnet3_input_node) = { .name = "vmxnet3-input", .sibling_of = "device-input", + .flags = VLIB_NODE_FLAG_TRACE_SUPPORTED, .format_trace = format_vmxnet3_input_trace, .type = VLIB_NODE_TYPE_INPUT, .state = VLIB_NODE_STATE_DISABLED, @@ -377,7 +492,6 @@ VLIB_REGISTER_NODE (vmxnet3_input_node) = { }; #endif -/* *INDENT-ON* */ /* * fd.io coding-style-patch-verification: ON