X-Git-Url: https://gerrit.fd.io/r/gitweb?a=blobdiff_plain;f=vnet%2Fvnet%2Fmap%2Fip6_map.c;h=2e38b0db62695e17c338137e5c10e6b663c03a16;hb=62f88d881306a1f628e3a25d3e3b00b8acf7ec5e;hp=208f45f521c68db002e4b931e016709942f86c98;hpb=52372b6c8e36401721a0533301e1ecdb30109424;p=vpp.git diff --git a/vnet/vnet/map/ip6_map.c b/vnet/vnet/map/ip6_map.c index 208f45f521c..2e38b0db626 100644 --- a/vnet/vnet/map/ip6_map.c +++ b/vnet/vnet/map/ip6_map.c @@ -16,7 +16,8 @@ #include "../ip/ip_frag.h" -enum ip6_map_next_e { +enum ip6_map_next_e +{ IP6_MAP_NEXT_IP4_LOOKUP, #ifdef MAP_SKIP_IP6_LOOKUP IP6_MAP_NEXT_IP4_REWRITE, @@ -31,20 +32,23 @@ enum ip6_map_next_e { IP6_MAP_N_NEXT, }; -enum ip6_map_ip6_reass_next_e { +enum ip6_map_ip6_reass_next_e +{ IP6_MAP_IP6_REASS_NEXT_IP6_MAP, IP6_MAP_IP6_REASS_NEXT_DROP, IP6_MAP_IP6_REASS_N_NEXT, }; -enum ip6_map_ip4_reass_next_e { +enum ip6_map_ip4_reass_next_e +{ IP6_MAP_IP4_REASS_NEXT_IP4_LOOKUP, IP6_MAP_IP4_REASS_NEXT_IP4_FRAGMENT, IP6_MAP_IP4_REASS_NEXT_DROP, IP6_MAP_IP4_REASS_N_NEXT, }; -enum ip6_icmp_relay_next_e { +enum ip6_icmp_relay_next_e +{ IP6_ICMP_RELAY_NEXT_IP4_LOOKUP, IP6_ICMP_RELAY_NEXT_DROP, IP6_ICMP_RELAY_N_NEXT, @@ -54,91 +58,113 @@ vlib_node_registration_t ip6_map_ip4_reass_node; vlib_node_registration_t ip6_map_ip6_reass_node; static vlib_node_registration_t ip6_map_icmp_relay_node; -typedef struct { +typedef struct +{ u32 map_domain_index; u16 port; u8 cached; } map_ip6_map_ip4_reass_trace_t; u8 * -format_ip6_map_ip4_reass_trace (u8 *s, va_list *args) +format_ip6_map_ip4_reass_trace (u8 * s, va_list * args) { - CLIB_UNUSED(vlib_main_t *vm) = va_arg (*args, vlib_main_t *); - CLIB_UNUSED(vlib_node_t *node) = va_arg (*args, vlib_node_t *); - map_ip6_map_ip4_reass_trace_t *t = va_arg (*args, map_ip6_map_ip4_reass_trace_t *); - return format(s, "MAP domain index: %d L4 port: %u Status: %s", t->map_domain_index, - t->port, t->cached?"cached":"forwarded"); + CLIB_UNUSED (vlib_main_t * vm) = va_arg (*args, vlib_main_t *); + CLIB_UNUSED (vlib_node_t * node) = va_arg (*args, vlib_node_t *); + map_ip6_map_ip4_reass_trace_t *t = + va_arg (*args, map_ip6_map_ip4_reass_trace_t *); + return format (s, "MAP domain index: %d L4 port: %u Status: %s", + t->map_domain_index, t->port, + t->cached ? "cached" : "forwarded"); } -typedef struct { +typedef struct +{ u16 offset; u16 frag_len; u8 out; } map_ip6_map_ip6_reass_trace_t; u8 * -format_ip6_map_ip6_reass_trace (u8 *s, va_list *args) +format_ip6_map_ip6_reass_trace (u8 * s, va_list * args) { - CLIB_UNUSED(vlib_main_t *vm) = va_arg (*args, vlib_main_t *); - CLIB_UNUSED(vlib_node_t *node) = va_arg (*args, vlib_node_t *); - map_ip6_map_ip6_reass_trace_t *t = va_arg (*args, map_ip6_map_ip6_reass_trace_t *); - return format(s, "Offset: %d Fragment length: %d Status: %s", t->offset, t->frag_len, t->out?"out":"in"); + CLIB_UNUSED (vlib_main_t * vm) = va_arg (*args, vlib_main_t *); + CLIB_UNUSED (vlib_node_t * node) = va_arg (*args, vlib_node_t *); + map_ip6_map_ip6_reass_trace_t *t = + va_arg (*args, map_ip6_map_ip6_reass_trace_t *); + return format (s, "Offset: %d Fragment length: %d Status: %s", t->offset, + t->frag_len, t->out ? "out" : "in"); } /* * ip6_map_sec_check */ static_always_inline bool -ip6_map_sec_check (map_domain_t *d, u16 port, ip4_header_t *ip4, ip6_header_t *ip6) +ip6_map_sec_check (map_domain_t * d, u16 port, ip4_header_t * ip4, + ip6_header_t * ip6) { - u16 sp4 = clib_net_to_host_u16(port); - u32 sa4 = clib_net_to_host_u32(ip4->src_address.as_u32); - u64 sal6 = map_get_pfx(d, sa4, sp4); - u64 sar6 = map_get_sfx(d, sa4, sp4); - - if (PREDICT_FALSE(sal6 != clib_net_to_host_u64(ip6->src_address.as_u64[0]) || - sar6 != clib_net_to_host_u64(ip6->src_address.as_u64[1]))) + u16 sp4 = clib_net_to_host_u16 (port); + u32 sa4 = clib_net_to_host_u32 (ip4->src_address.as_u32); + u64 sal6 = map_get_pfx (d, sa4, sp4); + u64 sar6 = map_get_sfx (d, sa4, sp4); + + if (PREDICT_FALSE + (sal6 != clib_net_to_host_u64 (ip6->src_address.as_u64[0]) + || sar6 != clib_net_to_host_u64 (ip6->src_address.as_u64[1]))) return (false); return (true); } static_always_inline void -ip6_map_security_check (map_domain_t *d, ip4_header_t *ip4, ip6_header_t *ip6, u32 *next, u8 *error) +ip6_map_security_check (map_domain_t * d, ip4_header_t * ip4, + ip6_header_t * ip6, u32 * next, u8 * error) { map_main_t *mm = &map_main; - if (d->ea_bits_len || d->rules) { - if (d->psid_length > 0) { - if (!ip4_is_fragment(ip4)) { - u16 port = ip4_map_get_port(ip4, MAP_SENDER); - if (port) { - if (mm->sec_check) - *error = ip6_map_sec_check(d, port, ip4, ip6) ? MAP_ERROR_NONE : MAP_ERROR_DECAP_SEC_CHECK; - } else { - *error = MAP_ERROR_BAD_PROTOCOL; + if (d->ea_bits_len || d->rules) + { + if (d->psid_length > 0) + { + if (!ip4_is_fragment (ip4)) + { + u16 port = ip4_map_get_port (ip4, MAP_SENDER); + if (port) + { + if (mm->sec_check) + *error = + ip6_map_sec_check (d, port, ip4, + ip6) ? MAP_ERROR_NONE : + MAP_ERROR_DECAP_SEC_CHECK; + } + else + { + *error = MAP_ERROR_BAD_PROTOCOL; + } + } + else + { + *next = mm->sec_check_frag ? IP6_MAP_NEXT_IP4_REASS : *next; + } } - } else { - *next = mm->sec_check_frag ? IP6_MAP_NEXT_IP4_REASS : *next; - } } - } } static_always_inline bool -ip6_map_ip4_lookup_bypass (vlib_buffer_t *p0, ip4_header_t *ip) +ip6_map_ip4_lookup_bypass (vlib_buffer_t * p0, ip4_header_t * ip) { #ifdef MAP_SKIP_IP6_LOOKUP map_main_t *mm = &map_main; u32 adj_index0 = mm->adj4_index; - if (adj_index0 > 0) { - ip_lookup_main_t *lm4 = &ip4_main.lookup_main; - ip_adjacency_t *adj = ip_get_adjacency(lm4, mm->adj4_index); - if (adj->n_adj > 1) { - u32 hash_c0 = ip4_compute_flow_hash(ip, IP_FLOW_HASH_DEFAULT); - adj_index0 += (hash_c0 & (adj->n_adj - 1)); + if (adj_index0 > 0) + { + ip_lookup_main_t *lm4 = &ip4_main.lookup_main; + ip_adjacency_t *adj = ip_get_adjacency (lm4, mm->adj4_index); + if (adj->n_adj > 1) + { + u32 hash_c0 = ip4_compute_flow_hash (ip, IP_FLOW_HASH_DEFAULT); + adj_index0 += (hash_c0 & (adj->n_adj - 1)); + } + vnet_buffer (p0)->ip.adj_index[VLIB_TX] = adj_index0; + return (true); } - vnet_buffer(p0)->ip.adj_index[VLIB_TX] = adj_index0; - return (true); - } #endif return (false); } @@ -147,376 +173,503 @@ ip6_map_ip4_lookup_bypass (vlib_buffer_t *p0, ip4_header_t *ip) * ip6_map */ static uword -ip6_map (vlib_main_t *vm, - vlib_node_runtime_t *node, - vlib_frame_t *frame) +ip6_map (vlib_main_t * vm, vlib_node_runtime_t * node, vlib_frame_t * frame) { u32 n_left_from, *from, next_index, *to_next, n_left_to_next; - vlib_node_runtime_t *error_node = vlib_node_get_runtime(vm, ip6_map_node.index); + vlib_node_runtime_t *error_node = + vlib_node_get_runtime (vm, ip6_map_node.index); map_main_t *mm = &map_main; vlib_combined_counter_main_t *cm = mm->domain_counters; - u32 cpu_index = os_get_cpu_number(); + u32 cpu_index = os_get_cpu_number (); - from = vlib_frame_vector_args(frame); + from = vlib_frame_vector_args (frame); n_left_from = frame->n_vectors; next_index = node->cached_next_index; - while (n_left_from > 0) { - vlib_get_next_frame(vm, node, next_index, to_next, n_left_to_next); - - /* Dual loop */ - while (n_left_from >= 4 && n_left_to_next >= 2) { - u32 pi0, pi1; - vlib_buffer_t *p0, *p1; - u8 error0 = MAP_ERROR_NONE; - u8 error1 = MAP_ERROR_NONE; - map_domain_t *d0 = 0, *d1 = 0; - ip4_header_t *ip40, *ip41; - ip6_header_t *ip60, *ip61; - u16 port0 = 0, port1 = 0; - u32 map_domain_index0 = ~0, map_domain_index1 = ~0; - u32 next0 = IP6_MAP_NEXT_IP4_LOOKUP; - u32 next1 = IP6_MAP_NEXT_IP4_LOOKUP; - - /* Prefetch next iteration. */ - { - vlib_buffer_t *p2, *p3; - - p2 = vlib_get_buffer(vm, from[2]); - p3 = vlib_get_buffer(vm, from[3]); - - vlib_prefetch_buffer_header(p2, LOAD); - vlib_prefetch_buffer_header(p3, LOAD); - - /* IPv6 + IPv4 header + 8 bytes of ULP */ - CLIB_PREFETCH(p2->data, 68, LOAD); - CLIB_PREFETCH(p3->data, 68, LOAD); - } - - pi0 = to_next[0] = from[0]; - pi1 = to_next[1] = from[1]; - from += 2; - n_left_from -= 2; - to_next +=2; - n_left_to_next -= 2; - - p0 = vlib_get_buffer(vm, pi0); - p1 = vlib_get_buffer(vm, pi1); - ip60 = vlib_buffer_get_current(p0); - ip61 = vlib_buffer_get_current(p1); - vlib_buffer_advance(p0, sizeof(ip6_header_t)); - vlib_buffer_advance(p1, sizeof(ip6_header_t)); - ip40 = vlib_buffer_get_current(p0); - ip41 = vlib_buffer_get_current(p1); - - /* - * Encapsulated IPv4 packet - * - IPv4 fragmented -> Pass to virtual reassembly unless security check disabled - * - Lookup/Rewrite or Fragment node in case of packet > MTU - * Fragmented IPv6 packet - * ICMP IPv6 packet - * - Error -> Pass to ICMPv6/ICMPv4 relay - * - Info -> Pass to IPv6 local - * Anything else -> drop - */ - if (PREDICT_TRUE(ip60->protocol == IP_PROTOCOL_IP_IN_IP && clib_net_to_host_u16(ip60->payload_length) > 20)) { - d0 = ip6_map_get_domain(vnet_buffer(p0)->ip.adj_index[VLIB_TX], (ip4_address_t *)&ip40->src_address.as_u32, - &map_domain_index0, &error0); - } else if (ip60->protocol == IP_PROTOCOL_ICMP6 && - clib_net_to_host_u16(ip60->payload_length) > sizeof(icmp46_header_t)) { - icmp46_header_t *icmp = (void *)(ip60 + 1); - next0 = (icmp->type == ICMP6_echo_request || icmp->type == ICMP6_echo_reply) ? - IP6_MAP_NEXT_IP6_LOCAL : IP6_MAP_NEXT_IP6_ICMP_RELAY; - } else if (ip60->protocol == IP_PROTOCOL_IPV6_FRAGMENTATION) { - next0 = IP6_MAP_NEXT_IP6_REASS; - } else { - error0 = MAP_ERROR_BAD_PROTOCOL; - } - if (PREDICT_TRUE(ip61->protocol == IP_PROTOCOL_IP_IN_IP && clib_net_to_host_u16(ip61->payload_length) > 20)) { - d1 = ip6_map_get_domain(vnet_buffer(p1)->ip.adj_index[VLIB_TX], (ip4_address_t *)&ip41->src_address.as_u32, - &map_domain_index1, &error1); - } else if (ip61->protocol == IP_PROTOCOL_ICMP6 && - clib_net_to_host_u16(ip61->payload_length) > sizeof(icmp46_header_t)) { - icmp46_header_t *icmp = (void *)(ip61 + 1); - next1 = (icmp->type == ICMP6_echo_request || icmp->type == ICMP6_echo_reply) ? - IP6_MAP_NEXT_IP6_LOCAL : IP6_MAP_NEXT_IP6_ICMP_RELAY; - } else if (ip61->protocol == IP_PROTOCOL_IPV6_FRAGMENTATION) { - next1 = IP6_MAP_NEXT_IP6_REASS; - } else { - error1 = MAP_ERROR_BAD_PROTOCOL; - } - - if (d0) { - /* MAP inbound security check */ - ip6_map_security_check(d0, ip40, ip60, &next0, &error0); - - if (PREDICT_TRUE(error0 == MAP_ERROR_NONE && - next0 == IP6_MAP_NEXT_IP4_LOOKUP)) { - if (PREDICT_FALSE(d0->mtu && (clib_host_to_net_u16(ip40->length) > d0->mtu))) { - vnet_buffer(p0)->ip_frag.header_offset = 0; - vnet_buffer(p0)->ip_frag.flags = 0; - vnet_buffer(p0)->ip_frag.next_index = IP4_FRAG_NEXT_IP4_LOOKUP; - vnet_buffer(p0)->ip_frag.mtu = d0->mtu; - next0 = IP6_MAP_NEXT_IP4_FRAGMENT; - } else { - next0 = ip6_map_ip4_lookup_bypass(p0, ip40) ? IP6_MAP_NEXT_IP4_REWRITE : next0; - } - vlib_increment_combined_counter(cm + MAP_DOMAIN_COUNTER_RX, cpu_index, map_domain_index0, 1, - clib_net_to_host_u16(ip40->length)); - } - } - if (d1) { - /* MAP inbound security check */ - ip6_map_security_check(d1, ip41, ip61, &next1, &error1); - - if (PREDICT_TRUE(error1 == MAP_ERROR_NONE && - next1 == IP6_MAP_NEXT_IP4_LOOKUP)) { - if (PREDICT_FALSE(d1->mtu && (clib_host_to_net_u16(ip41->length) > d1->mtu))) { - vnet_buffer(p1)->ip_frag.header_offset = 0; - vnet_buffer(p1)->ip_frag.flags = 0; - vnet_buffer(p1)->ip_frag.next_index = IP4_FRAG_NEXT_IP4_LOOKUP; - vnet_buffer(p1)->ip_frag.mtu = d0->mtu; - next1 = IP6_MAP_NEXT_IP4_FRAGMENT; - } else { - next1 = ip6_map_ip4_lookup_bypass(p1, ip41) ? IP6_MAP_NEXT_IP4_REWRITE : next1; + while (n_left_from > 0) + { + vlib_get_next_frame (vm, node, next_index, to_next, n_left_to_next); + + /* Dual loop */ + while (n_left_from >= 4 && n_left_to_next >= 2) + { + u32 pi0, pi1; + vlib_buffer_t *p0, *p1; + u8 error0 = MAP_ERROR_NONE; + u8 error1 = MAP_ERROR_NONE; + map_domain_t *d0 = 0, *d1 = 0; + ip4_header_t *ip40, *ip41; + ip6_header_t *ip60, *ip61; + u16 port0 = 0, port1 = 0; + u32 map_domain_index0 = ~0, map_domain_index1 = ~0; + u32 next0 = IP6_MAP_NEXT_IP4_LOOKUP; + u32 next1 = IP6_MAP_NEXT_IP4_LOOKUP; + + /* Prefetch next iteration. */ + { + vlib_buffer_t *p2, *p3; + + p2 = vlib_get_buffer (vm, from[2]); + p3 = vlib_get_buffer (vm, from[3]); + + vlib_prefetch_buffer_header (p2, LOAD); + vlib_prefetch_buffer_header (p3, LOAD); + + /* IPv6 + IPv4 header + 8 bytes of ULP */ + CLIB_PREFETCH (p2->data, 68, LOAD); + CLIB_PREFETCH (p3->data, 68, LOAD); } - vlib_increment_combined_counter(cm + MAP_DOMAIN_COUNTER_RX, cpu_index, map_domain_index1, 1, - clib_net_to_host_u16(ip41->length)); + + pi0 = to_next[0] = from[0]; + pi1 = to_next[1] = from[1]; + from += 2; + n_left_from -= 2; + to_next += 2; + n_left_to_next -= 2; + + p0 = vlib_get_buffer (vm, pi0); + p1 = vlib_get_buffer (vm, pi1); + ip60 = vlib_buffer_get_current (p0); + ip61 = vlib_buffer_get_current (p1); + vlib_buffer_advance (p0, sizeof (ip6_header_t)); + vlib_buffer_advance (p1, sizeof (ip6_header_t)); + ip40 = vlib_buffer_get_current (p0); + ip41 = vlib_buffer_get_current (p1); + + /* + * Encapsulated IPv4 packet + * - IPv4 fragmented -> Pass to virtual reassembly unless security check disabled + * - Lookup/Rewrite or Fragment node in case of packet > MTU + * Fragmented IPv6 packet + * ICMP IPv6 packet + * - Error -> Pass to ICMPv6/ICMPv4 relay + * - Info -> Pass to IPv6 local + * Anything else -> drop + */ + if (PREDICT_TRUE + (ip60->protocol == IP_PROTOCOL_IP_IN_IP + && clib_net_to_host_u16 (ip60->payload_length) > 20)) + { + d0 = + ip6_map_get_domain (vnet_buffer (p0)->ip.adj_index[VLIB_TX], + (ip4_address_t *) & ip40->src_address. + as_u32, &map_domain_index0, &error0); + } + else if (ip60->protocol == IP_PROTOCOL_ICMP6 && + clib_net_to_host_u16 (ip60->payload_length) > + sizeof (icmp46_header_t)) + { + icmp46_header_t *icmp = (void *) (ip60 + 1); + next0 = (icmp->type == ICMP6_echo_request + || icmp->type == + ICMP6_echo_reply) ? IP6_MAP_NEXT_IP6_LOCAL : + IP6_MAP_NEXT_IP6_ICMP_RELAY; + } + else if (ip60->protocol == IP_PROTOCOL_IPV6_FRAGMENTATION) + { + next0 = IP6_MAP_NEXT_IP6_REASS; + } + else + { + error0 = MAP_ERROR_BAD_PROTOCOL; + } + if (PREDICT_TRUE + (ip61->protocol == IP_PROTOCOL_IP_IN_IP + && clib_net_to_host_u16 (ip61->payload_length) > 20)) + { + d1 = + ip6_map_get_domain (vnet_buffer (p1)->ip.adj_index[VLIB_TX], + (ip4_address_t *) & ip41->src_address. + as_u32, &map_domain_index1, &error1); + } + else if (ip61->protocol == IP_PROTOCOL_ICMP6 && + clib_net_to_host_u16 (ip61->payload_length) > + sizeof (icmp46_header_t)) + { + icmp46_header_t *icmp = (void *) (ip61 + 1); + next1 = (icmp->type == ICMP6_echo_request + || icmp->type == + ICMP6_echo_reply) ? IP6_MAP_NEXT_IP6_LOCAL : + IP6_MAP_NEXT_IP6_ICMP_RELAY; + } + else if (ip61->protocol == IP_PROTOCOL_IPV6_FRAGMENTATION) + { + next1 = IP6_MAP_NEXT_IP6_REASS; + } + else + { + error1 = MAP_ERROR_BAD_PROTOCOL; + } + + if (d0) + { + /* MAP inbound security check */ + ip6_map_security_check (d0, ip40, ip60, &next0, &error0); + + if (PREDICT_TRUE (error0 == MAP_ERROR_NONE && + next0 == IP6_MAP_NEXT_IP4_LOOKUP)) + { + if (PREDICT_FALSE + (d0->mtu + && (clib_host_to_net_u16 (ip40->length) > d0->mtu))) + { + vnet_buffer (p0)->ip_frag.header_offset = 0; + vnet_buffer (p0)->ip_frag.flags = 0; + vnet_buffer (p0)->ip_frag.next_index = + IP4_FRAG_NEXT_IP4_LOOKUP; + vnet_buffer (p0)->ip_frag.mtu = d0->mtu; + next0 = IP6_MAP_NEXT_IP4_FRAGMENT; + } + else + { + next0 = + ip6_map_ip4_lookup_bypass (p0, + ip40) ? + IP6_MAP_NEXT_IP4_REWRITE : next0; + } + vlib_increment_combined_counter (cm + MAP_DOMAIN_COUNTER_RX, + cpu_index, + map_domain_index0, 1, + clib_net_to_host_u16 + (ip40->length)); + } + } + if (d1) + { + /* MAP inbound security check */ + ip6_map_security_check (d1, ip41, ip61, &next1, &error1); + + if (PREDICT_TRUE (error1 == MAP_ERROR_NONE && + next1 == IP6_MAP_NEXT_IP4_LOOKUP)) + { + if (PREDICT_FALSE + (d1->mtu + && (clib_host_to_net_u16 (ip41->length) > d1->mtu))) + { + vnet_buffer (p1)->ip_frag.header_offset = 0; + vnet_buffer (p1)->ip_frag.flags = 0; + vnet_buffer (p1)->ip_frag.next_index = + IP4_FRAG_NEXT_IP4_LOOKUP; + vnet_buffer (p1)->ip_frag.mtu = d1->mtu; + next1 = IP6_MAP_NEXT_IP4_FRAGMENT; + } + else + { + next1 = + ip6_map_ip4_lookup_bypass (p1, + ip41) ? + IP6_MAP_NEXT_IP4_REWRITE : next1; + } + vlib_increment_combined_counter (cm + MAP_DOMAIN_COUNTER_RX, + cpu_index, + map_domain_index1, 1, + clib_net_to_host_u16 + (ip41->length)); + } + } + + if (PREDICT_FALSE (p0->flags & VLIB_BUFFER_IS_TRACED)) + { + map_trace_t *tr = vlib_add_trace (vm, node, p0, sizeof (*tr)); + tr->map_domain_index = map_domain_index0; + tr->port = port0; + } + + if (PREDICT_FALSE (p1->flags & VLIB_BUFFER_IS_TRACED)) + { + map_trace_t *tr = vlib_add_trace (vm, node, p1, sizeof (*tr)); + tr->map_domain_index = map_domain_index1; + tr->port = port1; + } + + if (error0 == MAP_ERROR_DECAP_SEC_CHECK && mm->icmp6_enabled) + { + /* Set ICMP parameters */ + vlib_buffer_advance (p0, -sizeof (ip6_header_t)); + icmp6_error_set_vnet_buffer (p0, ICMP6_destination_unreachable, + ICMP6_destination_unreachable_source_address_failed_policy, + 0); + next0 = IP6_MAP_NEXT_ICMP; + } + else + { + next0 = (error0 == MAP_ERROR_NONE) ? next0 : IP6_MAP_NEXT_DROP; + } + + if (error1 == MAP_ERROR_DECAP_SEC_CHECK && mm->icmp6_enabled) + { + /* Set ICMP parameters */ + vlib_buffer_advance (p1, -sizeof (ip6_header_t)); + icmp6_error_set_vnet_buffer (p1, ICMP6_destination_unreachable, + ICMP6_destination_unreachable_source_address_failed_policy, + 0); + next1 = IP6_MAP_NEXT_ICMP; + } + else + { + next1 = (error1 == MAP_ERROR_NONE) ? next1 : IP6_MAP_NEXT_DROP; + } + + /* Reset packet */ + if (next0 == IP6_MAP_NEXT_IP6_LOCAL) + vlib_buffer_advance (p0, -sizeof (ip6_header_t)); + if (next1 == IP6_MAP_NEXT_IP6_LOCAL) + vlib_buffer_advance (p1, -sizeof (ip6_header_t)); + + p0->error = error_node->errors[error0]; + p1->error = error_node->errors[error1]; + vlib_validate_buffer_enqueue_x2 (vm, node, next_index, to_next, + n_left_to_next, pi0, pi1, next0, + next1); } - } - - if (PREDICT_FALSE(p0->flags & VLIB_BUFFER_IS_TRACED)) { - map_trace_t *tr = vlib_add_trace(vm, node, p0, sizeof(*tr)); - tr->map_domain_index = map_domain_index0; - tr->port = port0; - } - - if (PREDICT_FALSE(p1->flags & VLIB_BUFFER_IS_TRACED)) { - map_trace_t *tr = vlib_add_trace(vm, node, p1, sizeof(*tr)); - tr->map_domain_index = map_domain_index1; - tr->port = port1; - } - - if (error0 == MAP_ERROR_DECAP_SEC_CHECK && mm->icmp6_enabled) { - /* Set ICMP parameters */ - vlib_buffer_advance(p0, -sizeof(ip6_header_t)); - icmp6_error_set_vnet_buffer(p0, ICMP6_destination_unreachable, - ICMP6_destination_unreachable_source_address_failed_policy, 0); - next0 = IP6_MAP_NEXT_ICMP; - } else { - next0 = (error0 == MAP_ERROR_NONE) ? next0 : IP6_MAP_NEXT_DROP; - } - - if (error1 == MAP_ERROR_DECAP_SEC_CHECK && mm->icmp6_enabled) { - /* Set ICMP parameters */ - vlib_buffer_advance(p1, -sizeof(ip6_header_t)); - icmp6_error_set_vnet_buffer(p1, ICMP6_destination_unreachable, - ICMP6_destination_unreachable_source_address_failed_policy, 0); - next1 = IP6_MAP_NEXT_ICMP; - } else { - next1 = (error1 == MAP_ERROR_NONE) ? next1 : IP6_MAP_NEXT_DROP; - } - - /* Reset packet */ - if (next0 == IP6_MAP_NEXT_IP6_LOCAL) - vlib_buffer_advance(p0, -sizeof(ip6_header_t)); - if (next1 == IP6_MAP_NEXT_IP6_LOCAL) - vlib_buffer_advance(p1, -sizeof(ip6_header_t)); - - p0->error = error_node->errors[error0]; - p1->error = error_node->errors[error1]; - vlib_validate_buffer_enqueue_x2(vm, node, next_index, to_next, n_left_to_next, pi0, pi1, next0, next1); - } - /* Single loop */ - while (n_left_from > 0 && n_left_to_next > 0) { - u32 pi0; - vlib_buffer_t *p0; - u8 error0 = MAP_ERROR_NONE; - map_domain_t *d0 = 0; - ip4_header_t *ip40; - ip6_header_t *ip60; - i32 port0 = 0; - u32 map_domain_index0 = ~0; - u32 next0 = IP6_MAP_NEXT_IP4_LOOKUP; - - pi0 = to_next[0] = from[0]; - from += 1; - n_left_from -= 1; - to_next +=1; - n_left_to_next -= 1; - - p0 = vlib_get_buffer(vm, pi0); - ip60 = vlib_buffer_get_current(p0); - vlib_buffer_advance(p0, sizeof(ip6_header_t)); - ip40 = vlib_buffer_get_current(p0); - - /* - * Encapsulated IPv4 packet - * - IPv4 fragmented -> Pass to virtual reassembly unless security check disabled - * - Lookup/Rewrite or Fragment node in case of packet > MTU - * Fragmented IPv6 packet - * ICMP IPv6 packet - * - Error -> Pass to ICMPv6/ICMPv4 relay - * - Info -> Pass to IPv6 local - * Anything else -> drop - */ - if (PREDICT_TRUE(ip60->protocol == IP_PROTOCOL_IP_IN_IP && clib_net_to_host_u16(ip60->payload_length) > 20)) { - d0 = ip6_map_get_domain(vnet_buffer(p0)->ip.adj_index[VLIB_TX], (ip4_address_t *)&ip40->src_address.as_u32, - &map_domain_index0, &error0); - } else if (ip60->protocol == IP_PROTOCOL_ICMP6 && - clib_net_to_host_u16(ip60->payload_length) > sizeof(icmp46_header_t)) { - icmp46_header_t *icmp = (void *)(ip60 + 1); - next0 = (icmp->type == ICMP6_echo_request || icmp->type == ICMP6_echo_reply) ? - IP6_MAP_NEXT_IP6_LOCAL : IP6_MAP_NEXT_IP6_ICMP_RELAY; - } else if (ip60->protocol == IP_PROTOCOL_IPV6_FRAGMENTATION && - (((ip6_frag_hdr_t *)(ip60+1))->next_hdr == IP_PROTOCOL_IP_IN_IP)) { - next0 = IP6_MAP_NEXT_IP6_REASS; - } else { - error0 = MAP_ERROR_BAD_PROTOCOL; - } - - if (d0) { - /* MAP inbound security check */ - ip6_map_security_check(d0, ip40, ip60, &next0, &error0); - - if (PREDICT_TRUE(error0 == MAP_ERROR_NONE && - next0 == IP6_MAP_NEXT_IP4_LOOKUP)) { - if (PREDICT_FALSE(d0->mtu && (clib_host_to_net_u16(ip40->length) > d0->mtu))) { - vnet_buffer(p0)->ip_frag.header_offset = 0; - vnet_buffer(p0)->ip_frag.flags = 0; - vnet_buffer(p0)->ip_frag.next_index = IP4_FRAG_NEXT_IP4_LOOKUP; - vnet_buffer(p0)->ip_frag.mtu = d0->mtu; - next0 = IP6_MAP_NEXT_IP4_FRAGMENT; - } else { - next0 = ip6_map_ip4_lookup_bypass(p0, ip40) ? IP6_MAP_NEXT_IP4_REWRITE : next0; - } - vlib_increment_combined_counter(cm + MAP_DOMAIN_COUNTER_RX, cpu_index, map_domain_index0, 1, - clib_net_to_host_u16(ip40->length)); + /* Single loop */ + while (n_left_from > 0 && n_left_to_next > 0) + { + u32 pi0; + vlib_buffer_t *p0; + u8 error0 = MAP_ERROR_NONE; + map_domain_t *d0 = 0; + ip4_header_t *ip40; + ip6_header_t *ip60; + i32 port0 = 0; + u32 map_domain_index0 = ~0; + u32 next0 = IP6_MAP_NEXT_IP4_LOOKUP; + + pi0 = to_next[0] = from[0]; + from += 1; + n_left_from -= 1; + to_next += 1; + n_left_to_next -= 1; + + p0 = vlib_get_buffer (vm, pi0); + ip60 = vlib_buffer_get_current (p0); + vlib_buffer_advance (p0, sizeof (ip6_header_t)); + ip40 = vlib_buffer_get_current (p0); + + /* + * Encapsulated IPv4 packet + * - IPv4 fragmented -> Pass to virtual reassembly unless security check disabled + * - Lookup/Rewrite or Fragment node in case of packet > MTU + * Fragmented IPv6 packet + * ICMP IPv6 packet + * - Error -> Pass to ICMPv6/ICMPv4 relay + * - Info -> Pass to IPv6 local + * Anything else -> drop + */ + if (PREDICT_TRUE + (ip60->protocol == IP_PROTOCOL_IP_IN_IP + && clib_net_to_host_u16 (ip60->payload_length) > 20)) + { + d0 = + ip6_map_get_domain (vnet_buffer (p0)->ip.adj_index[VLIB_TX], + (ip4_address_t *) & ip40->src_address. + as_u32, &map_domain_index0, &error0); + } + else if (ip60->protocol == IP_PROTOCOL_ICMP6 && + clib_net_to_host_u16 (ip60->payload_length) > + sizeof (icmp46_header_t)) + { + icmp46_header_t *icmp = (void *) (ip60 + 1); + next0 = (icmp->type == ICMP6_echo_request + || icmp->type == + ICMP6_echo_reply) ? IP6_MAP_NEXT_IP6_LOCAL : + IP6_MAP_NEXT_IP6_ICMP_RELAY; + } + else if (ip60->protocol == IP_PROTOCOL_IPV6_FRAGMENTATION && + (((ip6_frag_hdr_t *) (ip60 + 1))->next_hdr == + IP_PROTOCOL_IP_IN_IP)) + { + next0 = IP6_MAP_NEXT_IP6_REASS; + } + else + { + error0 = MAP_ERROR_BAD_PROTOCOL; + } + + if (d0) + { + /* MAP inbound security check */ + ip6_map_security_check (d0, ip40, ip60, &next0, &error0); + + if (PREDICT_TRUE (error0 == MAP_ERROR_NONE && + next0 == IP6_MAP_NEXT_IP4_LOOKUP)) + { + if (PREDICT_FALSE + (d0->mtu + && (clib_host_to_net_u16 (ip40->length) > d0->mtu))) + { + vnet_buffer (p0)->ip_frag.header_offset = 0; + vnet_buffer (p0)->ip_frag.flags = 0; + vnet_buffer (p0)->ip_frag.next_index = + IP4_FRAG_NEXT_IP4_LOOKUP; + vnet_buffer (p0)->ip_frag.mtu = d0->mtu; + next0 = IP6_MAP_NEXT_IP4_FRAGMENT; + } + else + { + next0 = + ip6_map_ip4_lookup_bypass (p0, + ip40) ? + IP6_MAP_NEXT_IP4_REWRITE : next0; + } + vlib_increment_combined_counter (cm + MAP_DOMAIN_COUNTER_RX, + cpu_index, + map_domain_index0, 1, + clib_net_to_host_u16 + (ip40->length)); + } + } + + if (PREDICT_FALSE (p0->flags & VLIB_BUFFER_IS_TRACED)) + { + map_trace_t *tr = vlib_add_trace (vm, node, p0, sizeof (*tr)); + tr->map_domain_index = map_domain_index0; + tr->port = (u16) port0; + } + + if (mm->icmp6_enabled && + (error0 == MAP_ERROR_DECAP_SEC_CHECK + || error0 == MAP_ERROR_NO_DOMAIN)) + { + /* Set ICMP parameters */ + vlib_buffer_advance (p0, -sizeof (ip6_header_t)); + icmp6_error_set_vnet_buffer (p0, ICMP6_destination_unreachable, + ICMP6_destination_unreachable_source_address_failed_policy, + 0); + next0 = IP6_MAP_NEXT_ICMP; + } + else + { + next0 = (error0 == MAP_ERROR_NONE) ? next0 : IP6_MAP_NEXT_DROP; + } + + /* Reset packet */ + if (next0 == IP6_MAP_NEXT_IP6_LOCAL) + vlib_buffer_advance (p0, -sizeof (ip6_header_t)); + + p0->error = error_node->errors[error0]; + vlib_validate_buffer_enqueue_x1 (vm, node, next_index, to_next, + n_left_to_next, pi0, next0); } - } - - if (PREDICT_FALSE(p0->flags & VLIB_BUFFER_IS_TRACED)) { - map_trace_t *tr = vlib_add_trace(vm, node, p0, sizeof(*tr)); - tr->map_domain_index = map_domain_index0; - tr->port = (u16)port0; - } - - if (mm->icmp6_enabled && - (error0 == MAP_ERROR_DECAP_SEC_CHECK || error0 == MAP_ERROR_NO_DOMAIN)) { - /* Set ICMP parameters */ - vlib_buffer_advance(p0, -sizeof(ip6_header_t)); - icmp6_error_set_vnet_buffer(p0, ICMP6_destination_unreachable, - ICMP6_destination_unreachable_source_address_failed_policy, 0); - next0 = IP6_MAP_NEXT_ICMP; - } else { - next0 = (error0 == MAP_ERROR_NONE) ? next0 : IP6_MAP_NEXT_DROP; - } - - /* Reset packet */ - if (next0 == IP6_MAP_NEXT_IP6_LOCAL) - vlib_buffer_advance(p0, -sizeof(ip6_header_t)); - - p0->error = error_node->errors[error0]; - vlib_validate_buffer_enqueue_x1(vm, node, next_index, to_next, n_left_to_next, pi0, next0); + vlib_put_next_frame (vm, node, next_index, n_left_to_next); } - vlib_put_next_frame(vm, node, next_index, n_left_to_next); - } return frame->n_vectors; } static_always_inline void -ip6_map_ip6_reass_prepare(vlib_main_t *vm, vlib_node_runtime_t *node, map_ip6_reass_t *r, - u32 **fragments_ready, u32 **fragments_to_drop) +ip6_map_ip6_reass_prepare (vlib_main_t * vm, vlib_node_runtime_t * node, + map_ip6_reass_t * r, u32 ** fragments_ready, + u32 ** fragments_to_drop) { ip4_header_t *ip40; ip6_header_t *ip60; ip6_frag_hdr_t *frag0; vlib_buffer_t *p0; - if(!r->ip4_header.ip_version_and_header_length) + if (!r->ip4_header.ip_version_and_header_length) return; //The IP header is here, we need to check for packets //that can be forwarded int i; - for (i=0; ifragments[i].pi == ~0 || - ((!r->fragments[i].next_data_len) && (r->fragments[i].next_data_offset != (0xffff)))) - continue; - - p0 = vlib_get_buffer(vm, r->fragments[i].pi); - ip60 = vlib_buffer_get_current(p0); - frag0 = (ip6_frag_hdr_t *)(ip60 + 1); - ip40 = (ip4_header_t *)(frag0 + 1); - - if (ip6_frag_hdr_offset(frag0)) { - //Not first fragment, add the IPv4 header - memcpy(ip40, &r->ip4_header, 20); - } + for (i = 0; i < MAP_IP6_REASS_MAX_FRAGMENTS_PER_REASSEMBLY; i++) + { + if (r->fragments[i].pi == ~0 || + ((!r->fragments[i].next_data_len) + && (r->fragments[i].next_data_offset != (0xffff)))) + continue; + + p0 = vlib_get_buffer (vm, r->fragments[i].pi); + ip60 = vlib_buffer_get_current (p0); + frag0 = (ip6_frag_hdr_t *) (ip60 + 1); + ip40 = (ip4_header_t *) (frag0 + 1); + + if (ip6_frag_hdr_offset (frag0)) + { + //Not first fragment, add the IPv4 header + clib_memcpy (ip40, &r->ip4_header, 20); + } #ifdef MAP_IP6_REASS_COUNT_BYTES - r->forwarded += clib_net_to_host_u16(ip60->payload_length) - sizeof(*frag0); + r->forwarded += + clib_net_to_host_u16 (ip60->payload_length) - sizeof (*frag0); #endif - if (ip6_frag_hdr_more(frag0)) { - //Not last fragment, we copy end of next - memcpy(u8_ptr_add(ip60, p0->current_length), r->fragments[i].next_data, 20); - p0->current_length += 20; - ip60->payload_length = u16_net_add(ip60->payload_length, 20); - } - - if (!ip4_is_fragment(ip40)) { - ip40->fragment_id = frag_id_6to4(frag0->identification); - ip40->flags_and_fragment_offset = clib_host_to_net_u16(ip6_frag_hdr_offset(frag0)); - } else { - ip40->flags_and_fragment_offset = clib_host_to_net_u16(ip4_get_fragment_offset(ip40) + ip6_frag_hdr_offset(frag0)); - } - - if (ip6_frag_hdr_more(frag0)) - ip40->flags_and_fragment_offset |= clib_host_to_net_u16(IP4_HEADER_FLAG_MORE_FRAGMENTS); + if (ip6_frag_hdr_more (frag0)) + { + //Not last fragment, we copy end of next + clib_memcpy (u8_ptr_add (ip60, p0->current_length), + r->fragments[i].next_data, 20); + p0->current_length += 20; + ip60->payload_length = u16_net_add (ip60->payload_length, 20); + } - ip40->length = clib_host_to_net_u16(p0->current_length - sizeof(*ip60) - sizeof(*frag0)); - ip40->checksum = ip4_header_checksum(ip40); + if (!ip4_is_fragment (ip40)) + { + ip40->fragment_id = frag_id_6to4 (frag0->identification); + ip40->flags_and_fragment_offset = + clib_host_to_net_u16 (ip6_frag_hdr_offset (frag0)); + } + else + { + ip40->flags_and_fragment_offset = + clib_host_to_net_u16 (ip4_get_fragment_offset (ip40) + + ip6_frag_hdr_offset (frag0)); + } - if (PREDICT_FALSE(p0->flags & VLIB_BUFFER_IS_TRACED)) { - map_ip6_map_ip6_reass_trace_t *tr = vlib_add_trace(vm, node, p0, sizeof(*tr)); - tr->offset = ip4_get_fragment_offset(ip40); - tr->frag_len = clib_net_to_host_u16(ip40->length) - sizeof(*ip40); - tr->out = 1; - } + if (ip6_frag_hdr_more (frag0)) + ip40->flags_and_fragment_offset |= + clib_host_to_net_u16 (IP4_HEADER_FLAG_MORE_FRAGMENTS); + + ip40->length = + clib_host_to_net_u16 (p0->current_length - sizeof (*ip60) - + sizeof (*frag0)); + ip40->checksum = ip4_header_checksum (ip40); + + if (PREDICT_FALSE (p0->flags & VLIB_BUFFER_IS_TRACED)) + { + map_ip6_map_ip6_reass_trace_t *tr = + vlib_add_trace (vm, node, p0, sizeof (*tr)); + tr->offset = ip4_get_fragment_offset (ip40); + tr->frag_len = clib_net_to_host_u16 (ip40->length) - sizeof (*ip40); + tr->out = 1; + } - vec_add1(*fragments_ready, r->fragments[i].pi); - r->fragments[i].pi = ~0; - r->fragments[i].next_data_len = 0; - r->fragments[i].next_data_offset = 0; - map_main.ip6_reass_buffered_counter--; + vec_add1 (*fragments_ready, r->fragments[i].pi); + r->fragments[i].pi = ~0; + r->fragments[i].next_data_len = 0; + r->fragments[i].next_data_offset = 0; + map_main.ip6_reass_buffered_counter--; - //TODO: Best solution would be that ip6_map handles extension headers - // and ignores atomic fragment. But in the meantime, let's just copy the header. + //TODO: Best solution would be that ip6_map handles extension headers + // and ignores atomic fragment. But in the meantime, let's just copy the header. - u8 protocol = frag0->next_hdr; - memmove(u8_ptr_add(ip40, - sizeof(*ip60)), ip60, sizeof(*ip60)); - ((ip6_header_t *)u8_ptr_add(ip40, - sizeof(*ip60)))->protocol = protocol; - vlib_buffer_advance(p0, sizeof(*frag0)); - } + u8 protocol = frag0->next_hdr; + memmove (u8_ptr_add (ip40, -sizeof (*ip60)), ip60, sizeof (*ip60)); + ((ip6_header_t *) u8_ptr_add (ip40, -sizeof (*ip60)))->protocol = + protocol; + vlib_buffer_advance (p0, sizeof (*frag0)); + } } void -map_ip6_drop_pi(u32 pi) +map_ip6_drop_pi (u32 pi) { - vlib_main_t *vm = vlib_get_main(); - vlib_node_runtime_t *n = vlib_node_get_runtime(vm, ip6_map_ip6_reass_node.index); - vlib_set_next_frame_buffer(vm, n, IP6_MAP_IP6_REASS_NEXT_DROP, pi); + vlib_main_t *vm = vlib_get_main (); + vlib_node_runtime_t *n = + vlib_node_get_runtime (vm, ip6_map_ip6_reass_node.index); + vlib_set_next_frame_buffer (vm, n, IP6_MAP_IP6_REASS_NEXT_DROP, pi); } void -map_ip4_drop_pi(u32 pi) +map_ip4_drop_pi (u32 pi) { - vlib_main_t *vm = vlib_get_main(); - vlib_node_runtime_t *n = vlib_node_get_runtime(vm, ip6_map_ip4_reass_node.index); - vlib_set_next_frame_buffer(vm, n, IP6_MAP_IP4_REASS_NEXT_DROP, pi); + vlib_main_t *vm = vlib_get_main (); + vlib_node_runtime_t *n = + vlib_node_get_runtime (vm, ip6_map_ip4_reass_node.index); + vlib_set_next_frame_buffer (vm, n, IP6_MAP_IP4_REASS_NEXT_DROP, pi); } /* @@ -527,109 +680,139 @@ map_ip4_drop_pi(u32 pi) * have been forwarded. */ static uword -ip6_map_ip6_reass (vlib_main_t *vm, - vlib_node_runtime_t *node, - vlib_frame_t *frame) +ip6_map_ip6_reass (vlib_main_t * vm, + vlib_node_runtime_t * node, vlib_frame_t * frame) { u32 n_left_from, *from, next_index, *to_next, n_left_to_next; - vlib_node_runtime_t *error_node = vlib_node_get_runtime(vm, ip6_map_ip6_reass_node.index); + vlib_node_runtime_t *error_node = + vlib_node_get_runtime (vm, ip6_map_ip6_reass_node.index); u32 *fragments_to_drop = NULL; u32 *fragments_ready = NULL; - from = vlib_frame_vector_args(frame); + from = vlib_frame_vector_args (frame); n_left_from = frame->n_vectors; next_index = node->cached_next_index; - while (n_left_from > 0) { - vlib_get_next_frame(vm, node, next_index, to_next, n_left_to_next); - - /* Single loop */ - while (n_left_from > 0 && n_left_to_next > 0) { - u32 pi0; - vlib_buffer_t *p0; - u8 error0 = MAP_ERROR_NONE; - ip6_header_t *ip60; - ip6_frag_hdr_t *frag0; - u16 offset; - u16 next_offset; - u16 frag_len; - - pi0 = to_next[0] = from[0]; - from += 1; - n_left_from -= 1; - to_next +=1; - n_left_to_next -= 1; - - p0 = vlib_get_buffer(vm, pi0); - ip60 = vlib_buffer_get_current(p0); - frag0 = (ip6_frag_hdr_t *)(ip60 + 1); - offset = clib_host_to_net_u16(frag0->fragment_offset_and_more) & (~7); - frag_len = clib_net_to_host_u16(ip60->payload_length) - sizeof(*frag0); - next_offset = ip6_frag_hdr_more(frag0) ? (offset + frag_len) : (0xffff); - - //FIXME: Support other extension headers, maybe - - if (PREDICT_FALSE(p0->flags & VLIB_BUFFER_IS_TRACED)) { - map_ip6_map_ip6_reass_trace_t *tr = vlib_add_trace(vm, node, p0, sizeof(*tr)); - tr->offset = offset; - tr->frag_len = frag_len; - tr->out = 0; - } - - map_ip6_reass_lock(); - map_ip6_reass_t *r = map_ip6_reass_get(&ip60->src_address, &ip60->dst_address, - frag0->identification, frag0->next_hdr, &fragments_to_drop); - //FIXME: Use better error codes - if (PREDICT_FALSE(!r)) { - // Could not create a caching entry - error0 = MAP_ERROR_FRAGMENT_MEMORY; - } else if (PREDICT_FALSE((frag_len <= 20 && - (ip6_frag_hdr_more(frag0) || (!offset))))) { - //Very small fragment are restricted to the last one and - //can't be the first one - error0 = MAP_ERROR_FRAGMENT_MALFORMED; - } else if (map_ip6_reass_add_fragment(r, pi0, offset, next_offset, (u8 *)(frag0 + 1), frag_len)) { - map_ip6_reass_free(r, &fragments_to_drop); - error0 = MAP_ERROR_FRAGMENT_MEMORY; - } else { + while (n_left_from > 0) + { + vlib_get_next_frame (vm, node, next_index, to_next, n_left_to_next); + + /* Single loop */ + while (n_left_from > 0 && n_left_to_next > 0) + { + u32 pi0; + vlib_buffer_t *p0; + u8 error0 = MAP_ERROR_NONE; + ip6_header_t *ip60; + ip6_frag_hdr_t *frag0; + u16 offset; + u16 next_offset; + u16 frag_len; + + pi0 = to_next[0] = from[0]; + from += 1; + n_left_from -= 1; + to_next += 1; + n_left_to_next -= 1; + + p0 = vlib_get_buffer (vm, pi0); + ip60 = vlib_buffer_get_current (p0); + frag0 = (ip6_frag_hdr_t *) (ip60 + 1); + offset = + clib_host_to_net_u16 (frag0->fragment_offset_and_more) & (~7); + frag_len = + clib_net_to_host_u16 (ip60->payload_length) - sizeof (*frag0); + next_offset = + ip6_frag_hdr_more (frag0) ? (offset + frag_len) : (0xffff); + + //FIXME: Support other extension headers, maybe + + if (PREDICT_FALSE (p0->flags & VLIB_BUFFER_IS_TRACED)) + { + map_ip6_map_ip6_reass_trace_t *tr = + vlib_add_trace (vm, node, p0, sizeof (*tr)); + tr->offset = offset; + tr->frag_len = frag_len; + tr->out = 0; + } + + map_ip6_reass_lock (); + map_ip6_reass_t *r = + map_ip6_reass_get (&ip60->src_address, &ip60->dst_address, + frag0->identification, frag0->next_hdr, + &fragments_to_drop); + //FIXME: Use better error codes + if (PREDICT_FALSE (!r)) + { + // Could not create a caching entry + error0 = MAP_ERROR_FRAGMENT_MEMORY; + } + else if (PREDICT_FALSE ((frag_len <= 20 && + (ip6_frag_hdr_more (frag0) || (!offset))))) + { + //Very small fragment are restricted to the last one and + //can't be the first one + error0 = MAP_ERROR_FRAGMENT_MALFORMED; + } + else + if (map_ip6_reass_add_fragment + (r, pi0, offset, next_offset, (u8 *) (frag0 + 1), frag_len)) + { + map_ip6_reass_free (r, &fragments_to_drop); + error0 = MAP_ERROR_FRAGMENT_MEMORY; + } + else + { #ifdef MAP_IP6_REASS_COUNT_BYTES - if (!ip6_frag_hdr_more(frag0)) - r->expected_total = offset + frag_len; + if (!ip6_frag_hdr_more (frag0)) + r->expected_total = offset + frag_len; #endif - ip6_map_ip6_reass_prepare(vm, node, r, &fragments_ready, &fragments_to_drop); + ip6_map_ip6_reass_prepare (vm, node, r, &fragments_ready, + &fragments_to_drop); #ifdef MAP_IP6_REASS_COUNT_BYTES - if(r->forwarded >= r->expected_total) - map_ip6_reass_free(r, &fragments_to_drop); + if (r->forwarded >= r->expected_total) + map_ip6_reass_free (r, &fragments_to_drop); #endif - } - map_ip6_reass_unlock(); - - if (error0 == MAP_ERROR_NONE) { - if (frag_len > 20) { - //Dequeue the packet - n_left_to_next++; - to_next--; - } else { - //All data from that packet was copied no need to keep it, but this is not an error - p0->error = error_node->errors[MAP_ERROR_NONE]; - vlib_validate_buffer_enqueue_x1(vm, node, next_index, to_next, n_left_to_next, pi0, IP6_MAP_IP6_REASS_NEXT_DROP); - } - } else { - p0->error = error_node->errors[error0]; - vlib_validate_buffer_enqueue_x1(vm, node, next_index, to_next, n_left_to_next, pi0, IP6_MAP_IP6_REASS_NEXT_DROP); - } + } + map_ip6_reass_unlock (); + + if (error0 == MAP_ERROR_NONE) + { + if (frag_len > 20) + { + //Dequeue the packet + n_left_to_next++; + to_next--; + } + else + { + //All data from that packet was copied no need to keep it, but this is not an error + p0->error = error_node->errors[MAP_ERROR_NONE]; + vlib_validate_buffer_enqueue_x1 (vm, node, next_index, + to_next, n_left_to_next, + pi0, + IP6_MAP_IP6_REASS_NEXT_DROP); + } + } + else + { + p0->error = error_node->errors[error0]; + vlib_validate_buffer_enqueue_x1 (vm, node, next_index, to_next, + n_left_to_next, pi0, + IP6_MAP_IP6_REASS_NEXT_DROP); + } + } + vlib_put_next_frame (vm, node, next_index, n_left_to_next); } - vlib_put_next_frame(vm, node, next_index, n_left_to_next); - } - - map_send_all_to_node(vm, fragments_ready, node, - &error_node->errors[MAP_ERROR_NONE], - IP6_MAP_IP6_REASS_NEXT_IP6_MAP); - map_send_all_to_node(vm, fragments_to_drop, node, - &error_node->errors[MAP_ERROR_FRAGMENT_DROPPED], - IP6_MAP_IP6_REASS_NEXT_DROP); - - vec_free(fragments_to_drop); - vec_free(fragments_ready); + + map_send_all_to_node (vm, fragments_ready, node, + &error_node->errors[MAP_ERROR_NONE], + IP6_MAP_IP6_REASS_NEXT_IP6_MAP); + map_send_all_to_node (vm, fragments_to_drop, node, + &error_node->errors[MAP_ERROR_FRAGMENT_DROPPED], + IP6_MAP_IP6_REASS_NEXT_DROP); + + vec_free (fragments_to_drop); + vec_free (fragments_ready); return frame->n_vectors; } @@ -637,147 +820,195 @@ ip6_map_ip6_reass (vlib_main_t *vm, * ip6_ip4_virt_reass */ static uword -ip6_map_ip4_reass (vlib_main_t *vm, - vlib_node_runtime_t *node, - vlib_frame_t *frame) +ip6_map_ip4_reass (vlib_main_t * vm, + vlib_node_runtime_t * node, vlib_frame_t * frame) { u32 n_left_from, *from, next_index, *to_next, n_left_to_next; - vlib_node_runtime_t *error_node = vlib_node_get_runtime(vm, ip6_map_ip4_reass_node.index); + vlib_node_runtime_t *error_node = + vlib_node_get_runtime (vm, ip6_map_ip4_reass_node.index); map_main_t *mm = &map_main; vlib_combined_counter_main_t *cm = mm->domain_counters; - u32 cpu_index = os_get_cpu_number(); + u32 cpu_index = os_get_cpu_number (); u32 *fragments_to_drop = NULL; u32 *fragments_to_loopback = NULL; - from = vlib_frame_vector_args(frame); + from = vlib_frame_vector_args (frame); n_left_from = frame->n_vectors; next_index = node->cached_next_index; - while (n_left_from > 0) { - vlib_get_next_frame(vm, node, next_index, to_next, n_left_to_next); - - /* Single loop */ - while (n_left_from > 0 && n_left_to_next > 0) { - u32 pi0; - vlib_buffer_t *p0; - u8 error0 = MAP_ERROR_NONE; - map_domain_t *d0; - ip4_header_t *ip40; - ip6_header_t *ip60; - i32 port0 = 0; - u32 map_domain_index0 = ~0; - u32 next0 = IP6_MAP_IP4_REASS_NEXT_IP4_LOOKUP; - u8 cached = 0; - - pi0 = to_next[0] = from[0]; - from += 1; - n_left_from -= 1; - to_next +=1; - n_left_to_next -= 1; - - p0 = vlib_get_buffer(vm, pi0); - ip40 = vlib_buffer_get_current(p0); - ip60 = ((ip6_header_t *)ip40) - 1; - - d0 = ip6_map_get_domain(vnet_buffer(p0)->ip.adj_index[VLIB_TX], (ip4_address_t *)&ip40->src_address.as_u32, - &map_domain_index0, &error0); - - map_ip4_reass_lock(); - //This node only deals with fragmented ip4 - map_ip4_reass_t *r = map_ip4_reass_get(ip40->src_address.as_u32, ip40->dst_address.as_u32, - ip40->fragment_id, ip40->protocol, &fragments_to_drop); - if (PREDICT_FALSE(!r)) { - // Could not create a caching entry - error0 = MAP_ERROR_FRAGMENT_MEMORY; - } else if (PREDICT_TRUE(ip4_get_fragment_offset(ip40))) { - // This is a fragment - if (r->port >= 0) { - // We know the port already - port0 = r->port; - } else if (map_ip4_reass_add_fragment(r, pi0)) { - // Not enough space for caching - error0 = MAP_ERROR_FRAGMENT_MEMORY; - map_ip4_reass_free(r, &fragments_to_drop); - } else { - cached = 1; - } - } else if ((port0 = ip4_get_port(ip40, MAP_SENDER, p0->current_length)) < 0) { - // Could not find port from first fragment. Stop reassembling. - error0 = MAP_ERROR_BAD_PROTOCOL; - port0 = 0; - map_ip4_reass_free(r, &fragments_to_drop); - } else { - // Found port. Remember it and loopback saved fragments - r->port = port0; - map_ip4_reass_get_fragments(r, &fragments_to_loopback); - } + while (n_left_from > 0) + { + vlib_get_next_frame (vm, node, next_index, to_next, n_left_to_next); + + /* Single loop */ + while (n_left_from > 0 && n_left_to_next > 0) + { + u32 pi0; + vlib_buffer_t *p0; + u8 error0 = MAP_ERROR_NONE; + map_domain_t *d0; + ip4_header_t *ip40; + ip6_header_t *ip60; + i32 port0 = 0; + u32 map_domain_index0 = ~0; + u32 next0 = IP6_MAP_IP4_REASS_NEXT_IP4_LOOKUP; + u8 cached = 0; + + pi0 = to_next[0] = from[0]; + from += 1; + n_left_from -= 1; + to_next += 1; + n_left_to_next -= 1; + + p0 = vlib_get_buffer (vm, pi0); + ip40 = vlib_buffer_get_current (p0); + ip60 = ((ip6_header_t *) ip40) - 1; + + d0 = + ip6_map_get_domain (vnet_buffer (p0)->ip.adj_index[VLIB_TX], + (ip4_address_t *) & ip40->src_address.as_u32, + &map_domain_index0, &error0); + + map_ip4_reass_lock (); + //This node only deals with fragmented ip4 + map_ip4_reass_t *r = map_ip4_reass_get (ip40->src_address.as_u32, + ip40->dst_address.as_u32, + ip40->fragment_id, + ip40->protocol, + &fragments_to_drop); + if (PREDICT_FALSE (!r)) + { + // Could not create a caching entry + error0 = MAP_ERROR_FRAGMENT_MEMORY; + } + else if (PREDICT_TRUE (ip4_get_fragment_offset (ip40))) + { + // This is a fragment + if (r->port >= 0) + { + // We know the port already + port0 = r->port; + } + else if (map_ip4_reass_add_fragment (r, pi0)) + { + // Not enough space for caching + error0 = MAP_ERROR_FRAGMENT_MEMORY; + map_ip4_reass_free (r, &fragments_to_drop); + } + else + { + cached = 1; + } + } + else + if ((port0 = + ip4_get_port (ip40, MAP_SENDER, p0->current_length)) < 0) + { + // Could not find port from first fragment. Stop reassembling. + error0 = MAP_ERROR_BAD_PROTOCOL; + port0 = 0; + map_ip4_reass_free (r, &fragments_to_drop); + } + else + { + // Found port. Remember it and loopback saved fragments + r->port = port0; + map_ip4_reass_get_fragments (r, &fragments_to_loopback); + } #ifdef MAP_IP4_REASS_COUNT_BYTES - if (!cached && r) { - r->forwarded += clib_host_to_net_u16(ip40->length) - 20; - if (!ip4_get_fragment_more(ip40)) - r->expected_total = ip4_get_fragment_offset(ip40) * 8 + clib_host_to_net_u16(ip40->length) - 20; - if(r->forwarded >= r->expected_total) - map_ip4_reass_free(r, &fragments_to_drop); - } + if (!cached && r) + { + r->forwarded += clib_host_to_net_u16 (ip40->length) - 20; + if (!ip4_get_fragment_more (ip40)) + r->expected_total = + ip4_get_fragment_offset (ip40) * 8 + + clib_host_to_net_u16 (ip40->length) - 20; + if (r->forwarded >= r->expected_total) + map_ip4_reass_free (r, &fragments_to_drop); + } #endif - map_ip4_reass_unlock(); - - if(PREDICT_TRUE(error0 == MAP_ERROR_NONE)) - error0 = ip6_map_sec_check(d0, port0, ip40, ip60) ? MAP_ERROR_NONE : MAP_ERROR_DECAP_SEC_CHECK; - - if (PREDICT_FALSE(d0->mtu && (clib_host_to_net_u16(ip40->length) > d0->mtu) && - error0 == MAP_ERROR_NONE && !cached)) { - vnet_buffer(p0)->ip_frag.header_offset = 0; - vnet_buffer(p0)->ip_frag.flags = 0; - vnet_buffer(p0)->ip_frag.next_index = IP4_FRAG_NEXT_IP4_LOOKUP; - vnet_buffer(p0)->ip_frag.mtu = d0->mtu; - next0 = IP6_MAP_IP4_REASS_NEXT_IP4_FRAGMENT; - } - - if (PREDICT_FALSE(p0->flags & VLIB_BUFFER_IS_TRACED)) { - map_ip6_map_ip4_reass_trace_t *tr = vlib_add_trace(vm, node, p0, sizeof(*tr)); - tr->map_domain_index = map_domain_index0; - tr->port = port0; - tr->cached = cached; - } - - if (cached) { - //Dequeue the packet - n_left_to_next++; - to_next--; - } else { - if (error0 == MAP_ERROR_NONE) - vlib_increment_combined_counter(cm + MAP_DOMAIN_COUNTER_RX, cpu_index, map_domain_index0, 1, - clib_net_to_host_u16(ip40->length)); - next0 = (error0 == MAP_ERROR_NONE) ? next0 : IP6_MAP_IP4_REASS_NEXT_DROP; - p0->error = error_node->errors[error0]; - vlib_validate_buffer_enqueue_x1(vm, node, next_index, to_next, n_left_to_next, pi0, next0); - } - - //Loopback when we reach the end of the inpu vector - if(n_left_from == 0 && vec_len(fragments_to_loopback)) { - from = vlib_frame_vector_args(frame); - u32 len = vec_len(fragments_to_loopback); - if(len <= VLIB_FRAME_SIZE) { - memcpy(from, fragments_to_loopback, sizeof(u32)*len); - n_left_from = len; - vec_reset_length(fragments_to_loopback); - } else { - memcpy(from, fragments_to_loopback + (len - VLIB_FRAME_SIZE), sizeof(u32)*VLIB_FRAME_SIZE); - n_left_from = VLIB_FRAME_SIZE; - _vec_len(fragments_to_loopback) = len - VLIB_FRAME_SIZE; - } - } + map_ip4_reass_unlock (); + + if (PREDICT_TRUE (error0 == MAP_ERROR_NONE)) + error0 = + ip6_map_sec_check (d0, port0, ip40, + ip60) ? MAP_ERROR_NONE : + MAP_ERROR_DECAP_SEC_CHECK; + + if (PREDICT_FALSE + (d0->mtu && (clib_host_to_net_u16 (ip40->length) > d0->mtu) + && error0 == MAP_ERROR_NONE && !cached)) + { + vnet_buffer (p0)->ip_frag.header_offset = 0; + vnet_buffer (p0)->ip_frag.flags = 0; + vnet_buffer (p0)->ip_frag.next_index = IP4_FRAG_NEXT_IP4_LOOKUP; + vnet_buffer (p0)->ip_frag.mtu = d0->mtu; + next0 = IP6_MAP_IP4_REASS_NEXT_IP4_FRAGMENT; + } + + if (PREDICT_FALSE (p0->flags & VLIB_BUFFER_IS_TRACED)) + { + map_ip6_map_ip4_reass_trace_t *tr = + vlib_add_trace (vm, node, p0, sizeof (*tr)); + tr->map_domain_index = map_domain_index0; + tr->port = port0; + tr->cached = cached; + } + + if (cached) + { + //Dequeue the packet + n_left_to_next++; + to_next--; + } + else + { + if (error0 == MAP_ERROR_NONE) + vlib_increment_combined_counter (cm + MAP_DOMAIN_COUNTER_RX, + cpu_index, map_domain_index0, + 1, + clib_net_to_host_u16 + (ip40->length)); + next0 = + (error0 == + MAP_ERROR_NONE) ? next0 : IP6_MAP_IP4_REASS_NEXT_DROP; + p0->error = error_node->errors[error0]; + vlib_validate_buffer_enqueue_x1 (vm, node, next_index, to_next, + n_left_to_next, pi0, next0); + } + + //Loopback when we reach the end of the inpu vector + if (n_left_from == 0 && vec_len (fragments_to_loopback)) + { + from = vlib_frame_vector_args (frame); + u32 len = vec_len (fragments_to_loopback); + if (len <= VLIB_FRAME_SIZE) + { + clib_memcpy (from, fragments_to_loopback, + sizeof (u32) * len); + n_left_from = len; + vec_reset_length (fragments_to_loopback); + } + else + { + clib_memcpy (from, + fragments_to_loopback + (len - + VLIB_FRAME_SIZE), + sizeof (u32) * VLIB_FRAME_SIZE); + n_left_from = VLIB_FRAME_SIZE; + _vec_len (fragments_to_loopback) = len - VLIB_FRAME_SIZE; + } + } + } + vlib_put_next_frame (vm, node, next_index, n_left_to_next); } - vlib_put_next_frame(vm, node, next_index, n_left_to_next); - } - map_send_all_to_node(vm, fragments_to_drop, node, - &error_node->errors[MAP_ERROR_FRAGMENT_DROPPED], - IP6_MAP_IP4_REASS_NEXT_DROP); - - vec_free(fragments_to_drop); - vec_free(fragments_to_loopback); + map_send_all_to_node (vm, fragments_to_drop, node, + &error_node->errors[MAP_ERROR_FRAGMENT_DROPPED], + IP6_MAP_IP4_REASS_NEXT_DROP); + + vec_free (fragments_to_drop); + vec_free (fragments_to_loopback); return frame->n_vectors; } @@ -785,142 +1016,159 @@ ip6_map_ip4_reass (vlib_main_t *vm, * ip6_icmp_relay */ static uword -ip6_map_icmp_relay (vlib_main_t *vm, - vlib_node_runtime_t *node, - vlib_frame_t *frame) +ip6_map_icmp_relay (vlib_main_t * vm, + vlib_node_runtime_t * node, vlib_frame_t * frame) { u32 n_left_from, *from, next_index, *to_next, n_left_to_next; - vlib_node_runtime_t *error_node = vlib_node_get_runtime(vm, ip6_map_icmp_relay_node.index); + vlib_node_runtime_t *error_node = + vlib_node_get_runtime (vm, ip6_map_icmp_relay_node.index); map_main_t *mm = &map_main; - u32 cpu_index = os_get_cpu_number(); + u32 cpu_index = os_get_cpu_number (); u16 *fragment_ids, *fid; - from = vlib_frame_vector_args(frame); + from = vlib_frame_vector_args (frame); n_left_from = frame->n_vectors; next_index = node->cached_next_index; /* Get random fragment IDs for replies. */ - fid = fragment_ids = clib_random_buffer_get_data (&vm->random_buffer, n_left_from * sizeof (fragment_ids[0])); - - while (n_left_from > 0) { - vlib_get_next_frame(vm, node, next_index, to_next, n_left_to_next); - - /* Single loop */ - while (n_left_from > 0 && n_left_to_next > 0) { - u32 pi0; - vlib_buffer_t *p0; - u8 error0 = MAP_ERROR_NONE; - ip6_header_t *ip60; - u32 next0 = IP6_ICMP_RELAY_NEXT_IP4_LOOKUP; - u32 mtu; - - pi0 = to_next[0] = from[0]; - from += 1; - n_left_from -= 1; - to_next +=1; - n_left_to_next -= 1; - - p0 = vlib_get_buffer(vm, pi0); - ip60 = vlib_buffer_get_current(p0); - u16 tlen = clib_net_to_host_u16(ip60->payload_length); - - /* - * In: - * IPv6 header (40) - * ICMPv6 header (8) - * IPv6 header (40) - * Original IPv4 header / packet - * Out: - * New IPv4 header - * New ICMP header - * Original IPv4 header / packet - */ - - /* Need at least ICMP(8) + IPv6(40) + IPv4(20) + L4 header(8) */ - if (tlen < 76) { - error0 = MAP_ERROR_ICMP_RELAY; - goto error; - } - - icmp46_header_t *icmp60 = (icmp46_header_t *)(ip60 + 1); - ip6_header_t *inner_ip60 = (ip6_header_t *)(icmp60 + 2); - - if (inner_ip60->protocol != IP_PROTOCOL_IP_IN_IP) { - error0 = MAP_ERROR_ICMP_RELAY; - goto error; - } - - ip4_header_t *inner_ip40 = (ip4_header_t *)(inner_ip60 + 1); - vlib_buffer_advance(p0, 60); /* sizeof ( IPv6 + ICMP + IPv6 - IPv4 - ICMP ) */ - ip4_header_t *new_ip40 = vlib_buffer_get_current(p0); - icmp46_header_t *new_icmp40 = (icmp46_header_t *)(new_ip40 + 1); - - /* - * Relay according to RFC2473, section 8.3 - */ - switch (icmp60->type) { - case ICMP6_destination_unreachable: - case ICMP6_time_exceeded: - case ICMP6_parameter_problem: - /* Type 3 - destination unreachable, Code 1 - host unreachable */ - new_icmp40->type = ICMP4_destination_unreachable; - new_icmp40->code = ICMP4_destination_unreachable_destination_unreachable_host; - break; - - case ICMP6_packet_too_big: - /* Type 3 - destination unreachable, Code 4 - packet too big */ - /* Potential TODO: Adjust domain tunnel MTU based on the value received here */ - mtu = clib_net_to_host_u32(*((u32 *)(icmp60 + 1))); - - /* Check DF flag */ - if (!(inner_ip40->flags_and_fragment_offset & clib_host_to_net_u16(IP4_HEADER_FLAG_DONT_FRAGMENT))) { - error0 = MAP_ERROR_ICMP_RELAY; - goto error; + fid = fragment_ids = + clib_random_buffer_get_data (&vm->random_buffer, + n_left_from * sizeof (fragment_ids[0])); + + while (n_left_from > 0) + { + vlib_get_next_frame (vm, node, next_index, to_next, n_left_to_next); + + /* Single loop */ + while (n_left_from > 0 && n_left_to_next > 0) + { + u32 pi0; + vlib_buffer_t *p0; + u8 error0 = MAP_ERROR_NONE; + ip6_header_t *ip60; + u32 next0 = IP6_ICMP_RELAY_NEXT_IP4_LOOKUP; + u32 mtu; + + pi0 = to_next[0] = from[0]; + from += 1; + n_left_from -= 1; + to_next += 1; + n_left_to_next -= 1; + + p0 = vlib_get_buffer (vm, pi0); + ip60 = vlib_buffer_get_current (p0); + u16 tlen = clib_net_to_host_u16 (ip60->payload_length); + + /* + * In: + * IPv6 header (40) + * ICMPv6 header (8) + * IPv6 header (40) + * Original IPv4 header / packet + * Out: + * New IPv4 header + * New ICMP header + * Original IPv4 header / packet + */ + + /* Need at least ICMP(8) + IPv6(40) + IPv4(20) + L4 header(8) */ + if (tlen < 76) + { + error0 = MAP_ERROR_ICMP_RELAY; + goto error; + } + + icmp46_header_t *icmp60 = (icmp46_header_t *) (ip60 + 1); + ip6_header_t *inner_ip60 = (ip6_header_t *) (icmp60 + 2); + + if (inner_ip60->protocol != IP_PROTOCOL_IP_IN_IP) + { + error0 = MAP_ERROR_ICMP_RELAY; + goto error; + } + + ip4_header_t *inner_ip40 = (ip4_header_t *) (inner_ip60 + 1); + vlib_buffer_advance (p0, 60); /* sizeof ( IPv6 + ICMP + IPv6 - IPv4 - ICMP ) */ + ip4_header_t *new_ip40 = vlib_buffer_get_current (p0); + icmp46_header_t *new_icmp40 = (icmp46_header_t *) (new_ip40 + 1); + + /* + * Relay according to RFC2473, section 8.3 + */ + switch (icmp60->type) + { + case ICMP6_destination_unreachable: + case ICMP6_time_exceeded: + case ICMP6_parameter_problem: + /* Type 3 - destination unreachable, Code 1 - host unreachable */ + new_icmp40->type = ICMP4_destination_unreachable; + new_icmp40->code = + ICMP4_destination_unreachable_destination_unreachable_host; + break; + + case ICMP6_packet_too_big: + /* Type 3 - destination unreachable, Code 4 - packet too big */ + /* Potential TODO: Adjust domain tunnel MTU based on the value received here */ + mtu = clib_net_to_host_u32 (*((u32 *) (icmp60 + 1))); + + /* Check DF flag */ + if (! + (inner_ip40->flags_and_fragment_offset & + clib_host_to_net_u16 (IP4_HEADER_FLAG_DONT_FRAGMENT))) + { + error0 = MAP_ERROR_ICMP_RELAY; + goto error; + } + + new_icmp40->type = ICMP4_destination_unreachable; + new_icmp40->code = + ICMP4_destination_unreachable_fragmentation_needed_and_dont_fragment_set; + *((u32 *) (new_icmp40 + 1)) = + clib_host_to_net_u32 (mtu < 1280 ? 1280 : mtu); + break; + + default: + error0 = MAP_ERROR_ICMP_RELAY; + break; + } + + /* + * Ensure the total ICMP packet is no longer than 576 bytes (RFC1812) + */ + new_ip40->ip_version_and_header_length = 0x45; + new_ip40->tos = 0; + u16 nlen = (tlen - 20) > 576 ? 576 : tlen - 20; + new_ip40->length = clib_host_to_net_u16 (nlen); + new_ip40->fragment_id = fid[0]; + fid++; + new_ip40->ttl = 64; + new_ip40->protocol = IP_PROTOCOL_ICMP; + new_ip40->src_address = mm->icmp4_src_address; + new_ip40->dst_address = inner_ip40->src_address; + new_ip40->checksum = ip4_header_checksum (new_ip40); + + new_icmp40->checksum = 0; + ip_csum_t sum = ip_incremental_checksum (0, new_icmp40, nlen - 20); + new_icmp40->checksum = ~ip_csum_fold (sum); + + vlib_increment_simple_counter (&mm->icmp_relayed, cpu_index, 0, 1); + + error: + if (PREDICT_FALSE (p0->flags & VLIB_BUFFER_IS_TRACED)) + { + map_trace_t *tr = vlib_add_trace (vm, node, p0, sizeof (*tr)); + tr->map_domain_index = 0; + tr->port = 0; + } + + next0 = + (error0 == MAP_ERROR_NONE) ? next0 : IP6_ICMP_RELAY_NEXT_DROP; + p0->error = error_node->errors[error0]; + vlib_validate_buffer_enqueue_x1 (vm, node, next_index, to_next, + n_left_to_next, pi0, next0); } - - new_icmp40->type = ICMP4_destination_unreachable; - new_icmp40->code = ICMP4_destination_unreachable_fragmentation_needed_and_dont_fragment_set; - *((u32 *)(new_icmp40 + 1)) = clib_host_to_net_u32(mtu < 1280 ? 1280 : mtu); - break; - - default: - error0 = MAP_ERROR_ICMP_RELAY; - break; - } - - /* - * Ensure the total ICMP packet is no longer than 576 bytes (RFC1812) - */ - new_ip40->ip_version_and_header_length = 0x45; - new_ip40->tos = 0; - u16 nlen = (tlen - 20) > 576 ? 576 : tlen - 20; - new_ip40->length = clib_host_to_net_u16(nlen); - new_ip40->fragment_id = fid[0]; fid++; - new_ip40->ttl = 64; - new_ip40->protocol = IP_PROTOCOL_ICMP; - new_ip40->src_address = mm->icmp4_src_address; - new_ip40->dst_address = inner_ip40->src_address; - new_ip40->checksum = ip4_header_checksum(new_ip40); - - new_icmp40->checksum = 0; - ip_csum_t sum = ip_incremental_checksum(0, new_icmp40, nlen - 20); - new_icmp40->checksum = ~ip_csum_fold(sum); - - vlib_increment_simple_counter(&mm->icmp_relayed, cpu_index, 0, 1); - - error: - if (PREDICT_FALSE(p0->flags & VLIB_BUFFER_IS_TRACED)) { - map_trace_t *tr = vlib_add_trace(vm, node, p0, sizeof(*tr)); - tr->map_domain_index = 0; - tr->port = 0; - } - - next0 = (error0 == MAP_ERROR_NONE) ? next0 : IP6_ICMP_RELAY_NEXT_DROP; - p0->error = error_node->errors[error0]; - vlib_validate_buffer_enqueue_x1(vm, node, next_index, to_next, n_left_to_next, pi0, next0); + vlib_put_next_frame (vm, node, next_index, n_left_to_next); } - vlib_put_next_frame(vm, node, next_index, n_left_to_next); - } return frame->n_vectors; @@ -932,6 +1180,7 @@ static char *map_error_strings[] = { #undef _ }; +/* *INDENT-OFF* */ VLIB_REGISTER_NODE(ip6_map_node) = { .function = ip6_map, .name = "ip6-map", @@ -957,7 +1206,9 @@ VLIB_REGISTER_NODE(ip6_map_node) = { [IP6_MAP_NEXT_ICMP] = "ip6-icmp-error", }, }; +/* *INDENT-ON* */ +/* *INDENT-OFF* */ VLIB_REGISTER_NODE(ip6_map_ip6_reass_node) = { .function = ip6_map_ip6_reass, .name = "ip6-map-ip6-reass", @@ -972,7 +1223,9 @@ VLIB_REGISTER_NODE(ip6_map_ip6_reass_node) = { [IP6_MAP_IP6_REASS_NEXT_DROP] = "error-drop", }, }; +/* *INDENT-ON* */ +/* *INDENT-OFF* */ VLIB_REGISTER_NODE(ip6_map_ip4_reass_node) = { .function = ip6_map_ip4_reass, .name = "ip6-map-ip4-reass", @@ -988,7 +1241,9 @@ VLIB_REGISTER_NODE(ip6_map_ip4_reass_node) = { [IP6_MAP_IP4_REASS_NEXT_DROP] = "error-drop", }, }; +/* *INDENT-ON* */ +/* *INDENT-OFF* */ VLIB_REGISTER_NODE(ip6_map_icmp_relay_node, static) = { .function = ip6_map_icmp_relay, .name = "ip6-map-icmp-relay", @@ -1003,3 +1258,12 @@ VLIB_REGISTER_NODE(ip6_map_icmp_relay_node, static) = { [IP6_ICMP_RELAY_NEXT_DROP] = "error-drop", }, }; +/* *INDENT-ON* */ + +/* + * fd.io coding-style-patch-verification: ON + * + * Local Variables: + * eval: (c-set-style "gnu") + * End: + */