X-Git-Url: https://gerrit.fd.io/r/gitweb?a=blobdiff_plain;f=src%2Fvnet%2Fip%2Freass%2Fip6_sv_reass.c;h=e1493c95e54080910c780256a0f0e81f6b8b77fe;hb=8bea589cfe0fca1a6f560e16ca66a4cf199041a2;hp=c5c84e5f93a6f55956ba38e4308ee900c612ff29;hpb=b2c31b685fd2cf28436ca32bc93e23eb24c74878;p=vpp.git diff --git a/src/vnet/ip/reass/ip6_sv_reass.c b/src/vnet/ip/reass/ip6_sv_reass.c index c5c84e5f93a..e1493c95e54 100644 --- a/src/vnet/ip/reass/ip6_sv_reass.c +++ b/src/vnet/ip/reass/ip6_sv_reass.c @@ -26,6 +26,7 @@ #include #include #include +#include #define MSEC_PER_SEC 1000 #define IP6_SV_REASS_TIMEOUT_DEFAULT_MS 100 @@ -40,6 +41,7 @@ typedef enum IP6_SV_REASS_RC_TOO_MANY_FRAGMENTS, IP6_SV_REASS_RC_INTERNAL_ERROR, IP6_SV_REASS_RC_UNSUPP_IP_PROTO, + IP6_SV_REASS_RC_INVALID_FRAG_LEN, } ip6_sv_reass_rc_t; typedef struct @@ -50,7 +52,7 @@ typedef struct { ip6_address_t src; ip6_address_t dst; - u32 xx_id; + u32 fib_index; u32 frag_id; u8 unused[7]; u8 proto; @@ -214,7 +216,7 @@ format_ip6_sv_reass_trace (u8 * s, va_list * args) clib_net_to_host_u16 (t->l4_dst_port)); break; case REASS_PASSTHROUGH: - s = format (s, "[not-fragmented]"); + s = format (s, "[not fragmented or atomic fragment]"); break; } return s; @@ -222,12 +224,18 @@ format_ip6_sv_reass_trace (u8 * s, va_list * args) static void ip6_sv_reass_add_trace (vlib_main_t * vm, vlib_node_runtime_t * node, - ip6_sv_reass_main_t * rm, ip6_sv_reass_t * reass, u32 bi, ip6_sv_reass_trace_operation_e action, u32 ip_proto, u16 l4_src_port, u16 l4_dst_port) { vlib_buffer_t *b = vlib_get_buffer (vm, bi); + if (pool_is_free_index + (vm->trace_main.trace_buffer_pool, vlib_buffer_get_trace_index (b))) + { + // this buffer's trace is gone + b->flags &= ~VLIB_BUFFER_IS_TRACED; + return; + } ip6_sv_reass_trace_t *t = vlib_add_trace (vm, node, b, sizeof (t[0])); if (reass) { @@ -296,15 +304,15 @@ ip6_sv_reass_init (ip6_sv_reass_t * reass) } always_inline ip6_sv_reass_t * -ip6_sv_reass_find_or_create (vlib_main_t * vm, vlib_node_runtime_t * node, - ip6_sv_reass_main_t * rm, - ip6_sv_reass_per_thread_t * rt, - ip6_sv_reass_kv_t * kv, u32 * icmp_bi, - u8 * do_handoff) +ip6_sv_reass_find_or_create (vlib_main_t *vm, ip6_sv_reass_main_t *rm, + ip6_sv_reass_per_thread_t *rt, + ip6_sv_reass_kv_t *kv, u8 *do_handoff) { ip6_sv_reass_t *reass = NULL; f64 now = vlib_time_now (vm); +again: + if (!clib_bihash_search_48_8 (&rm->hash, &kv->kv, &kv->kv)) { if (vm->thread_index != kv->v.thread_index) @@ -364,20 +372,23 @@ ip6_sv_reass_find_or_create (vlib_main_t * vm, vlib_node_runtime_t * node, kv->v.thread_index = vm->thread_index; reass->last_heard = now; - if (clib_bihash_add_del_48_8 (&rm->hash, &kv->kv, 1)) + int rv = clib_bihash_add_del_48_8 (&rm->hash, &kv->kv, 2); + if (rv) { ip6_sv_reass_free (vm, rm, rt, reass); reass = NULL; + // if other worker created a context already work with the other copy + if (-2 == rv) + goto again; } return reass; } always_inline ip6_sv_reass_rc_t -ip6_sv_reass_update (vlib_main_t * vm, vlib_node_runtime_t * node, - ip6_sv_reass_main_t * rm, ip6_sv_reass_per_thread_t * rt, - ip6_sv_reass_t * reass, u32 bi0, - ip6_frag_hdr_t * frag_hdr) +ip6_sv_reass_update (vlib_main_t *vm, vlib_node_runtime_t *node, + ip6_sv_reass_main_t *rm, ip6_sv_reass_t *reass, u32 bi0, + ip6_frag_hdr_t *frag_hdr) { vlib_buffer_t *fb = vlib_get_buffer (vm, bi0); vnet_buffer_opaque_t *fvnb = vnet_buffer (fb); @@ -396,6 +407,10 @@ ip6_sv_reass_update (vlib_main_t * vm, vlib_node_runtime_t * node, u32 fragment_length = vlib_buffer_length_in_chain (vm, fb) - (fvnb->ip.reass.ip6_frag_hdr_offset + sizeof (*frag_hdr)); + if (0 == fragment_length) + { + return IP6_SV_REASS_RC_INVALID_FRAG_LEN; + } u32 fragment_last = fvnb->ip.reass.fragment_last = fragment_first + fragment_length - 1; fvnb->ip.reass.range_first = fragment_first; @@ -414,7 +429,7 @@ ip6_sv_reass_update (vlib_main_t * vm, vlib_node_runtime_t * node, vlib_buffer_t *b0 = vlib_get_buffer (vm, bi0); if (PREDICT_FALSE (b0->flags & VLIB_BUFFER_IS_TRACED)) { - ip6_sv_reass_add_trace (vm, node, rm, reass, bi0, REASS_FINISH, + ip6_sv_reass_add_trace (vm, node, reass, bi0, REASS_FINISH, reass->ip_proto, reass->l4_src_port, reass->l4_dst_port); } @@ -424,9 +439,9 @@ ip6_sv_reass_update (vlib_main_t * vm, vlib_node_runtime_t * node, { if (PREDICT_FALSE (fb->flags & VLIB_BUFFER_IS_TRACED)) { - ip6_sv_reass_add_trace (vm, node, rm, reass, bi0, - REASS_FRAGMENT_CACHE, reass->ip_proto, - reass->l4_src_port, reass->l4_dst_port); + ip6_sv_reass_add_trace (vm, node, reass, bi0, REASS_FRAGMENT_CACHE, + reass->ip_proto, reass->l4_src_port, + reass->l4_dst_port); } if (vec_len (reass->cached_buffers) > rm->max_reass_len) { @@ -437,22 +452,18 @@ ip6_sv_reass_update (vlib_main_t * vm, vlib_node_runtime_t * node, } always_inline bool -ip6_sv_reass_verify_upper_layer_present (vlib_node_runtime_t * node, - vlib_buffer_t * b, - ip6_frag_hdr_t * frag_hdr) +ip6_sv_reass_verify_upper_layer_present (vlib_node_runtime_t *node, + vlib_buffer_t *b, + ip6_ext_hdr_chain_t *hc) { - ip6_ext_header_t *tmp = (ip6_ext_header_t *) frag_hdr; - while (ip6_ext_hdr (tmp->next_hdr)) + int nh = hc->eh[hc->length - 1].protocol; + /* Checking to see if it's a terminating header */ + if (ip6_ext_hdr (nh)) { - tmp = ip6_ext_next_header (tmp); - } - if (IP_PROTOCOL_IP6_NONXT == tmp->next_hdr) - { - icmp6_error_set_vnet_buffer (b, ICMP6_parameter_problem, - ICMP6_parameter_problem_first_fragment_has_incomplete_header_chain, - 0); + icmp6_error_set_vnet_buffer ( + b, ICMP6_parameter_problem, + ICMP6_parameter_problem_first_fragment_has_incomplete_header_chain, 0); b->error = node->errors[IP6_ERROR_REASS_MISSING_UPPER]; - return false; } return true; @@ -460,7 +471,6 @@ ip6_sv_reass_verify_upper_layer_present (vlib_node_runtime_t * node, always_inline bool ip6_sv_reass_verify_fragment_multiple_8 (vlib_main_t * vm, - vlib_node_runtime_t * node, vlib_buffer_t * b, ip6_frag_hdr_t * frag_hdr) { @@ -482,7 +492,6 @@ ip6_sv_reass_verify_fragment_multiple_8 (vlib_main_t * vm, always_inline bool ip6_sv_reass_verify_packet_size_lt_64k (vlib_main_t * vm, - vlib_node_runtime_t * node, vlib_buffer_t * b, ip6_frag_hdr_t * frag_hdr) { @@ -527,24 +536,31 @@ ip6_sv_reassembly_inline (vlib_main_t * vm, vlib_buffer_t *b0; u32 next0 = IP6_SV_REASSEMBLY_NEXT_DROP; u32 error0 = IP6_ERROR_NONE; - u32 icmp_bi = ~0; bi0 = from[0]; b0 = vlib_get_buffer (vm, bi0); ip6_header_t *ip0 = vlib_buffer_get_current (b0); - ip6_frag_hdr_t *frag_hdr = NULL; - ip6_ext_header_t *prev_hdr; - if (ip6_ext_hdr (ip0->protocol)) + ip6_frag_hdr_t *frag_hdr; + ip6_ext_hdr_chain_t hdr_chain; + bool is_atomic_fragment = false; + + int res = ip6_ext_header_walk ( + b0, ip0, IP_PROTOCOL_IPV6_FRAGMENTATION, &hdr_chain); + if (res >= 0 && + hdr_chain.eh[res].protocol == IP_PROTOCOL_IPV6_FRAGMENTATION) { frag_hdr = - ip6_ext_header_find (vm, b0, ip0, - IP_PROTOCOL_IPV6_FRAGMENTATION, - &prev_hdr); + ip6_ext_next_header_offset (ip0, hdr_chain.eh[res].offset); + is_atomic_fragment = (0 == ip6_frag_hdr_offset (frag_hdr) && + !ip6_frag_hdr_more (frag_hdr)); } - if (!frag_hdr) + + if (res < 0 || + hdr_chain.eh[res].protocol != IP_PROTOCOL_IPV6_FRAGMENTATION || + is_atomic_fragment) { - // this is a regular packet - no fragmentation + // this is a regular unfragmented packet or an atomic fragment if (!ip6_get_port (vm, b0, ip0, b0->current_length, &(vnet_buffer (b0)->ip.reass.ip_proto), @@ -563,32 +579,30 @@ ip6_sv_reassembly_inline (vlib_main_t * vm, next0 = IP6_SV_REASSEMBLY_NEXT_INPUT; if (PREDICT_FALSE (b0->flags & VLIB_BUFFER_IS_TRACED)) { - ip6_sv_reass_add_trace (vm, node, rm, NULL, bi0, - REASS_PASSTHROUGH, - vnet_buffer (b0)->ip.reass.ip_proto, - vnet_buffer (b0)->ip. - reass.l4_src_port, - vnet_buffer (b0)->ip. - reass.l4_dst_port); + ip6_sv_reass_add_trace ( + vm, node, NULL, bi0, REASS_PASSTHROUGH, + vnet_buffer (b0)->ip.reass.ip_proto, + vnet_buffer (b0)->ip.reass.l4_src_port, + vnet_buffer (b0)->ip.reass.l4_dst_port); } goto packet_enqueue; } + vnet_buffer (b0)->ip.reass.ip6_frag_hdr_offset = - (u8 *) frag_hdr - (u8 *) ip0; + hdr_chain.eh[res].offset; + if (0 == ip6_frag_hdr_offset (frag_hdr)) { // first fragment - verify upper-layer is present - if (!ip6_sv_reass_verify_upper_layer_present - (node, b0, frag_hdr)) + if (!ip6_sv_reass_verify_upper_layer_present (node, b0, + &hdr_chain)) { next0 = IP6_SV_REASSEMBLY_NEXT_ICMP_ERROR; goto packet_enqueue; } } - if (!ip6_sv_reass_verify_fragment_multiple_8 - (vm, node, b0, frag_hdr) - || !ip6_sv_reass_verify_packet_size_lt_64k (vm, node, b0, - frag_hdr)) + if (!ip6_sv_reass_verify_fragment_multiple_8 (vm, b0, frag_hdr) || + !ip6_sv_reass_verify_packet_size_lt_64k (vm, b0, frag_hdr)) { next0 = IP6_SV_REASSEMBLY_NEXT_ICMP_ERROR; goto packet_enqueue; @@ -608,8 +622,7 @@ ip6_sv_reassembly_inline (vlib_main_t * vm, kv.k.as_u64[5] = ip0->protocol; ip6_sv_reass_t *reass = - ip6_sv_reass_find_or_create (vm, node, rm, rt, &kv, &icmp_bi, - &do_handoff); + ip6_sv_reass_find_or_create (vm, rm, rt, &kv, &do_handoff); if (PREDICT_FALSE (do_handoff)) { @@ -643,41 +656,37 @@ ip6_sv_reassembly_inline (vlib_main_t * vm, next0 = IP6_SV_REASSEMBLY_NEXT_INPUT; if (PREDICT_FALSE (b0->flags & VLIB_BUFFER_IS_TRACED)) { - ip6_sv_reass_add_trace (vm, node, rm, reass, bi0, - REASS_FRAGMENT_FORWARD, - reass->ip_proto, - reass->l4_src_port, - reass->l4_dst_port); + ip6_sv_reass_add_trace ( + vm, node, reass, bi0, REASS_FRAGMENT_FORWARD, + reass->ip_proto, reass->l4_src_port, reass->l4_dst_port); } goto packet_enqueue; } - switch (ip6_sv_reass_update - (vm, node, rm, rt, reass, bi0, frag_hdr)) + u32 counter = ~0; + switch (ip6_sv_reass_update (vm, node, rm, reass, bi0, frag_hdr)) { case IP6_SV_REASS_RC_OK: /* nothing to do here */ break; case IP6_SV_REASS_RC_TOO_MANY_FRAGMENTS: - vlib_node_increment_counter (vm, node->node_index, - IP6_ERROR_REASS_FRAGMENT_CHAIN_TOO_LONG, - 1); - ip6_sv_reass_free (vm, rm, rt, reass); - goto next_packet; + counter = IP6_ERROR_REASS_FRAGMENT_CHAIN_TOO_LONG; break; case IP6_SV_REASS_RC_UNSUPP_IP_PROTO: - vlib_node_increment_counter (vm, node->node_index, - IP6_ERROR_REASS_UNSUPP_IP_PROTO, - 1); - ip6_sv_reass_free (vm, rm, rt, reass); - goto next_packet; + counter = IP6_ERROR_REASS_UNSUPP_IP_PROTO; break; case IP6_SV_REASS_RC_INTERNAL_ERROR: - vlib_node_increment_counter (vm, node->node_index, - IP6_ERROR_REASS_INTERNAL_ERROR, 1); + counter = IP6_ERROR_REASS_INTERNAL_ERROR; + break; + case IP6_SV_REASS_RC_INVALID_FRAG_LEN: + counter = IP6_ERROR_REASS_INVALID_FRAG_LEN; + break; + } + if (~0 != counter) + { + vlib_node_increment_counter (vm, node->node_index, counter, 1); ip6_sv_reass_free (vm, rm, rt, reass); goto next_packet; - break; } if (reass->is_complete) @@ -717,17 +726,16 @@ ip6_sv_reassembly_inline (vlib_main_t * vm, vnet_buffer (b0)->ip.reass.l4_dst_port = reass->l4_dst_port; if (PREDICT_FALSE (b0->flags & VLIB_BUFFER_IS_TRACED)) { - ip6_sv_reass_add_trace (vm, node, rm, reass, bi0, - REASS_FRAGMENT_FORWARD, - reass->ip_proto, - reass->l4_src_port, - reass->l4_dst_port); + ip6_sv_reass_add_trace ( + vm, node, reass, bi0, REASS_FRAGMENT_FORWARD, + reass->ip_proto, reass->l4_src_port, reass->l4_dst_port); } vlib_validate_buffer_enqueue_x1 (vm, node, next_index, to_next, n_left_to_next, bi0, next0); } - _vec_len (reass->cached_buffers) = 0; // buffers are owned by frame now + vec_set_len (reass->cached_buffers, + 0); // buffers are owned by frame now } goto next_packet; @@ -743,17 +751,6 @@ ip6_sv_reassembly_inline (vlib_main_t * vm, vlib_validate_buffer_enqueue_x1 (vm, node, next_index, to_next, n_left_to_next, bi0, next0); - if (~0 != icmp_bi) - { - next0 = IP6_SV_REASSEMBLY_NEXT_ICMP_ERROR; - to_next[0] = icmp_bi; - to_next += 1; - n_left_to_next -= 1; - vlib_validate_buffer_enqueue_x1 (vm, node, next_index, to_next, - n_left_to_next, icmp_bi, - next0); - } - next_packet: from += 1; n_left_from -= 1; @@ -976,8 +973,6 @@ ip6_sv_reass_init_function (vlib_main_t * vm) if ((error = vlib_call_init_function (vm, ip_main_init))) return error; - ip6_register_protocol (IP_PROTOCOL_IPV6_FRAGMENTATION, - ip6_sv_reass_node.index); rm->fq_index = vlib_frame_queue_main_init (ip6_sv_reass_node.index, 0); rm->fq_feature_index = @@ -992,8 +987,9 @@ VLIB_INIT_FUNCTION (ip6_sv_reass_init_function); #endif /* CLIB_MARCH_VARIANT */ static uword -ip6_sv_reass_walk_expired (vlib_main_t * vm, - vlib_node_runtime_t * node, vlib_frame_t * f) +ip6_sv_reass_walk_expired (vlib_main_t *vm, + CLIB_UNUSED (vlib_node_runtime_t *node), + CLIB_UNUSED (vlib_frame_t *f)) { ip6_sv_reass_main_t *rm = &ip6_sv_reass_main; uword event_type, *event_data = 0; @@ -1007,10 +1003,11 @@ ip6_sv_reass_walk_expired (vlib_main_t * vm, switch (event_type) { - case ~0: /* no events => timeout */ - /* nothing to do here */ - break; + case ~0: + /* no events => timeout */ + /* fallthrough */ case IP6_EVENT_CONFIG_CHANGED: + /* nothing to do here */ break; default: clib_warning ("BUG: event type 0x%wx", event_type); @@ -1054,7 +1051,7 @@ ip6_sv_reass_walk_expired (vlib_main_t * vm, vec_free (pool_indexes_to_free); if (event_data) { - _vec_len (event_data) = 0; + vec_set_len (event_data, 0); } } @@ -1078,9 +1075,10 @@ static u8 * format_ip6_sv_reass_key (u8 * s, va_list * args) { ip6_sv_reass_key_t *key = va_arg (*args, ip6_sv_reass_key_t *); - s = format (s, "xx_id: %u, src: %U, dst: %U, frag_id: %u, proto: %u", - key->xx_id, format_ip6_address, &key->src, format_ip6_address, - &key->dst, clib_net_to_host_u16 (key->frag_id), key->proto); + s = + format (s, "fib_index: %u, src: %U, dst: %U, frag_id: %u, proto: %u", + key->fib_index, format_ip6_address, &key->src, format_ip6_address, + &key->dst, clib_net_to_host_u16 (key->frag_id), key->proto); return s; } @@ -1152,6 +1150,10 @@ show_ip6_sv_reass (vlib_main_t * vm, unformat_input_t * input, vlib_cli_output (vm, "Maximum configured concurrent shallow virtual IP6 reassemblies per worker-thread: %lu\n", (long unsigned) rm->max_reass_n); + vlib_cli_output (vm, + "Maximum configured amount of fragments per shallow " + "virtual IP6 reassembly: %lu\n", + (long unsigned) rm->max_reass_len); vlib_cli_output (vm, "Maximum configured shallow virtual IP6 reassembly timeout: %lums\n", (long unsigned) rm->timeout_ms); @@ -1256,9 +1258,8 @@ ip6_sv_reassembly_handoff_inline (vlib_main_t * vm, ti += 1; b += 1; } - n_enq = - vlib_buffer_enqueue_to_thread (vm, fq_index, from, thread_indices, - frame->n_vectors, 1); + n_enq = vlib_buffer_enqueue_to_thread (vm, node, fq_index, from, + thread_indices, frame->n_vectors, 1); if (n_enq < frame->n_vectors) vlib_node_increment_counter (vm, node->node_index,