X-Git-Url: https://gerrit.fd.io/r/gitweb?a=blobdiff_plain;f=src%2Fplugins%2Fmap%2Fip4_map_t.c;h=21d17d7791097212488cb750ba494e28012fdfe1;hb=fc7344f;hp=ab4b46c916f3232411ae4a7b4e9ac391861ab2e0;hpb=381e9a90748bb659f56081123052e3e95501a4b4;p=vpp.git diff --git a/src/plugins/map/ip4_map_t.c b/src/plugins/map/ip4_map_t.c index ab4b46c916f..21d17d77910 100644 --- a/src/plugins/map/ip4_map_t.c +++ b/src/plugins/map/ip4_map_t.c @@ -17,8 +17,6 @@ #include #include -#define IP4_MAP_T_DUAL_LOOP 1 - typedef enum { IP4_MAPT_NEXT_MAPT_TCP_UDP, @@ -102,33 +100,19 @@ ip4_map_fragment_get_port (ip4_header_t * ip4) typedef struct { map_domain_t *d; - u16 id; + u16 recv_port; } icmp_to_icmp6_ctx_t; static int ip4_to_ip6_set_icmp_cb (ip4_header_t * ip4, ip6_header_t * ip6, void *arg) { icmp_to_icmp6_ctx_t *ctx = arg; - map_main_t *mm = &map_main; - if (mm->is_ce) - { - ip6->src_address.as_u64[0] = - map_get_pfx_net (ctx->d, ip4->src_address.as_u32, ctx->id); - ip6->src_address.as_u64[1] = - map_get_sfx_net (ctx->d, ip4->src_address.as_u32, ctx->id); - ip4_map_t_embedded_address (ctx->d, &ip6->dst_address, - &ip4->dst_address); - } - else - { - ip4_map_t_embedded_address (ctx->d, &ip6->src_address, - &ip4->src_address); - ip6->dst_address.as_u64[0] = - map_get_pfx_net (ctx->d, ip4->dst_address.as_u32, ctx->id); - ip6->dst_address.as_u64[1] = - map_get_sfx_net (ctx->d, ip4->dst_address.as_u32, ctx->id); - } + ip4_map_t_embedded_address (ctx->d, &ip6->src_address, &ip4->src_address); + ip6->dst_address.as_u64[0] = + map_get_pfx_net (ctx->d, ip4->dst_address.as_u32, ctx->recv_port); + ip6->dst_address.as_u64[1] = + map_get_sfx_net (ctx->d, ip4->dst_address.as_u32, ctx->recv_port); return 0; } @@ -138,30 +122,14 @@ ip4_to_ip6_set_inner_icmp_cb (ip4_header_t * ip4, ip6_header_t * ip6, void *arg) { icmp_to_icmp6_ctx_t *ctx = arg; - map_main_t *mm = &map_main; - if (mm->is_ce) - { - //Note that the destination address is within the domain - //while the source address is the one outside the domain - ip4_map_t_embedded_address (ctx->d, &ip6->src_address, - &ip4->src_address); - ip6->dst_address.as_u64[0] = - map_get_pfx_net (ctx->d, ip4->dst_address.as_u32, ctx->id); - ip6->dst_address.as_u64[1] = - map_get_sfx_net (ctx->d, ip4->dst_address.as_u32, ctx->id); - } - else - { - //Note that the source address is within the domain - //while the destination address is the one outside the domain - ip4_map_t_embedded_address (ctx->d, &ip6->dst_address, - &ip4->dst_address); - ip6->src_address.as_u64[0] = - map_get_pfx_net (ctx->d, ip4->src_address.as_u32, ctx->id); - ip6->src_address.as_u64[1] = - map_get_sfx_net (ctx->d, ip4->src_address.as_u32, ctx->id); - } + //Note that the source address is within the domain + //while the destination address is the one outside the domain + ip4_map_t_embedded_address (ctx->d, &ip6->dst_address, &ip4->dst_address); + ip6->src_address.as_u64[0] = + map_get_pfx_net (ctx->d, ip4->src_address.as_u32, ctx->recv_port); + ip6->src_address.as_u64[1] = + map_get_sfx_net (ctx->d, ip4->src_address.as_u32, ctx->recv_port); return 0; } @@ -177,7 +145,7 @@ ip4_map_t_icmp (vlib_main_t * vm, n_left_from = frame->n_vectors; next_index = node->cached_next_index; vlib_combined_counter_main_t *cm = map_main.domain_counters; - u32 thread_index = vlib_get_thread_index (); + u32 thread_index = vm->thread_index; while (n_left_from > 0) { @@ -193,7 +161,6 @@ ip4_map_t_icmp (vlib_main_t * vm, u16 len0; icmp_to_icmp6_ctx_t ctx0; ip4_header_t *ip40; - icmp46_header_t *icmp0; next0 = IP4_MAPT_ICMP_NEXT_IP6_LOOKUP; pi0 = to_next[0] = from[0]; @@ -213,11 +180,9 @@ ip4_map_t_icmp (vlib_main_t * vm, vnet_buffer (p0)->map_t.map_domain_index); ip40 = vlib_buffer_get_current (p0); - icmp0 = (icmp46_header_t *) (ip40 + 1); - - ctx0.id = ip4_get_port (ip40, icmp0->type == ICMP6_echo_request); + ctx0.recv_port = ip4_get_port (ip40, 1); ctx0.d = d0; - if (ctx0.id == 0) + if (ctx0.recv_port == 0) { // In case of 1:1 mapping, we don't care about the port if (!(d0->ea_bits_len == 0 && d0->rules)) @@ -237,7 +202,6 @@ ip4_map_t_icmp (vlib_main_t * vm, if (vnet_buffer (p0)->map_t.mtu < p0->current_length) { - vnet_buffer (p0)->ip_frag.header_offset = 0; vnet_buffer (p0)->ip_frag.mtu = vnet_buffer (p0)->map_t.mtu; vnet_buffer (p0)->ip_frag.next_index = IP6_FRAG_NEXT_IP6_LOOKUP; next0 = IP4_MAPT_ICMP_NEXT_IP6_FRAG; @@ -265,10 +229,44 @@ ip4_map_t_icmp (vlib_main_t * vm, return frame->n_vectors; } -static int -ip4_to_ip6_set_cb (ip4_header_t * ip4, ip6_header_t * ip6, void *ctx) +/* + * Translate fragmented IPv4 UDP/TCP packet to IPv6. + */ +always_inline int +map_ip4_to_ip6_fragmented (vlib_buffer_t * p, + ip4_mapt_pseudo_header_t * pheader) { - ip4_mapt_pseudo_header_t *pheader = ctx; + ip4_header_t *ip4; + ip6_header_t *ip6; + ip6_frag_hdr_t *frag; + + ip4 = vlib_buffer_get_current (p); + frag = (ip6_frag_hdr_t *) u8_ptr_add (ip4, sizeof (*ip4) - sizeof (*frag)); + ip6 = + (ip6_header_t *) u8_ptr_add (ip4, + sizeof (*ip4) - sizeof (*frag) - + sizeof (*ip6)); + vlib_buffer_advance (p, sizeof (*ip4) - sizeof (*ip6) - sizeof (*frag)); + + //We know that the protocol was one of ICMP, TCP or UDP + //because the first fragment was found and cached + frag->next_hdr = + (ip4->protocol == IP_PROTOCOL_ICMP) ? IP_PROTOCOL_ICMP6 : ip4->protocol; + frag->identification = frag_id_4to6 (ip4->fragment_id); + frag->rsv = 0; + frag->fragment_offset_and_more = + ip6_frag_hdr_offset_and_more (ip4_get_fragment_offset (ip4), + clib_net_to_host_u16 + (ip4->flags_and_fragment_offset) & + IP4_HEADER_FLAG_MORE_FRAGMENTS); + + ip6->ip_version_traffic_class_and_flow_label = + clib_host_to_net_u32 ((6 << 28) + (ip4->tos << 20)); + ip6->payload_length = + clib_host_to_net_u16 (clib_net_to_host_u16 (ip4->length) - + sizeof (*ip4) + sizeof (*frag)); + ip6->hop_limit = ip4->ttl; + ip6->protocol = IP_PROTOCOL_IPV6_FRAGMENTATION; ip6->dst_address.as_u64[0] = pheader->daddr.as_u64[0]; ip6->dst_address.as_u64[1] = pheader->daddr.as_u64[1]; @@ -313,7 +311,7 @@ ip4_map_t_fragmented (vlib_main_t * vm, pheader0 = vlib_buffer_get_current (p0); vlib_buffer_advance (p0, sizeof (*pheader0)); - if (ip4_to_ip6_fragmented (p0, ip4_to_ip6_set_cb, pheader0)) + if (map_ip4_to_ip6_fragmented (p0, pheader0)) { p0->error = error_node->errors[MAP_ERROR_FRAGMENT_DROPPED]; next0 = IP4_MAPT_FRAGMENTED_NEXT_DROP; @@ -322,7 +320,6 @@ ip4_map_t_fragmented (vlib_main_t * vm, { if (vnet_buffer (p0)->map_t.mtu < p0->current_length) { - vnet_buffer (p0)->ip_frag.header_offset = 0; vnet_buffer (p0)->ip_frag.mtu = vnet_buffer (p0)->map_t.mtu; vnet_buffer (p0)->ip_frag.next_index = IP6_FRAG_NEXT_IP6_LOOKUP; @@ -339,6 +336,110 @@ ip4_map_t_fragmented (vlib_main_t * vm, return frame->n_vectors; } +/* + * Translate IPv4 UDP/TCP packet to IPv6. + */ +always_inline int +map_ip4_to_ip6_tcp_udp (vlib_buffer_t * p, ip4_mapt_pseudo_header_t * pheader) +{ + map_main_t *mm = &map_main; + ip4_header_t *ip4; + ip6_header_t *ip6; + ip_csum_t csum; + u16 *checksum; + ip6_frag_hdr_t *frag; + u32 frag_id; + ip4_address_t old_src, old_dst; + + ip4 = vlib_buffer_get_current (p); + + if (ip4->protocol == IP_PROTOCOL_UDP) + { + udp_header_t *udp = ip4_next_header (ip4); + checksum = &udp->checksum; + + /* + * UDP checksum is optional over IPv4 but mandatory for IPv6 We + * do not check udp->length sanity but use our safe computed + * value instead + */ + if (PREDICT_FALSE (!*checksum)) + { + u16 udp_len = clib_host_to_net_u16 (ip4->length) - sizeof (*ip4); + csum = ip_incremental_checksum (0, udp, udp_len); + csum = ip_csum_with_carry (csum, clib_host_to_net_u16 (udp_len)); + csum = + ip_csum_with_carry (csum, clib_host_to_net_u16 (IP_PROTOCOL_UDP)); + csum = ip_csum_with_carry (csum, *((u64 *) (&ip4->src_address))); + *checksum = ~ip_csum_fold (csum); + } + } + else + { + tcp_header_t *tcp = ip4_next_header (ip4); + if (mm->tcp_mss > 0) + { + csum = tcp->checksum; + map_mss_clamping (tcp, &csum, mm->tcp_mss); + tcp->checksum = ip_csum_fold (csum); + } + checksum = &tcp->checksum; + } + + old_src.as_u32 = ip4->src_address.as_u32; + old_dst.as_u32 = ip4->dst_address.as_u32; + + /* Deal with fragmented packets */ + if (PREDICT_FALSE (ip4->flags_and_fragment_offset & + clib_host_to_net_u16 (IP4_HEADER_FLAG_MORE_FRAGMENTS))) + { + ip6 = + (ip6_header_t *) u8_ptr_add (ip4, + sizeof (*ip4) - sizeof (*ip6) - + sizeof (*frag)); + frag = + (ip6_frag_hdr_t *) u8_ptr_add (ip4, sizeof (*ip4) - sizeof (*frag)); + frag_id = frag_id_4to6 (ip4->fragment_id); + vlib_buffer_advance (p, sizeof (*ip4) - sizeof (*ip6) - sizeof (*frag)); + } + else + { + ip6 = (ip6_header_t *) (((u8 *) ip4) + sizeof (*ip4) - sizeof (*ip6)); + vlib_buffer_advance (p, sizeof (*ip4) - sizeof (*ip6)); + frag = NULL; + } + + ip6->ip_version_traffic_class_and_flow_label = + clib_host_to_net_u32 ((6 << 28) + (ip4->tos << 20)); + ip6->payload_length = u16_net_add (ip4->length, -sizeof (*ip4)); + ip6->hop_limit = ip4->ttl; + ip6->protocol = ip4->protocol; + if (PREDICT_FALSE (frag != NULL)) + { + frag->next_hdr = ip6->protocol; + frag->identification = frag_id; + frag->rsv = 0; + frag->fragment_offset_and_more = ip6_frag_hdr_offset_and_more (0, 1); + ip6->protocol = IP_PROTOCOL_IPV6_FRAGMENTATION; + ip6->payload_length = u16_net_add (ip6->payload_length, sizeof (*frag)); + } + + ip6->dst_address.as_u64[0] = pheader->daddr.as_u64[0]; + ip6->dst_address.as_u64[1] = pheader->daddr.as_u64[1]; + ip6->src_address.as_u64[0] = pheader->saddr.as_u64[0]; + ip6->src_address.as_u64[1] = pheader->saddr.as_u64[1]; + + csum = ip_csum_sub_even (*checksum, old_src.as_u32); + csum = ip_csum_sub_even (csum, old_dst.as_u32); + csum = ip_csum_add_even (csum, ip6->src_address.as_u64[0]); + csum = ip_csum_add_even (csum, ip6->src_address.as_u64[1]); + csum = ip_csum_add_even (csum, ip6->dst_address.as_u64[0]); + csum = ip_csum_add_even (csum, ip6->dst_address.as_u64[1]); + *checksum = ip_csum_fold (csum); + + return 0; +} + static uword ip4_map_t_tcp_udp (vlib_main_t * vm, vlib_node_runtime_t * node, vlib_frame_t * frame) @@ -355,74 +456,6 @@ ip4_map_t_tcp_udp (vlib_main_t * vm, { vlib_get_next_frame (vm, node, next_index, to_next, n_left_to_next); -#ifdef IP4_MAP_T_DUAL_LOOP - while (n_left_from >= 4 && n_left_to_next >= 2) - { - u32 pi0, pi1; - vlib_buffer_t *p0, *p1; - ip4_mapt_pseudo_header_t *pheader0, *pheader1; - ip4_mapt_tcp_udp_next_t next0, next1; - - pi0 = to_next[0] = from[0]; - pi1 = to_next[1] = from[1]; - from += 2; - n_left_from -= 2; - to_next += 2; - n_left_to_next -= 2; - - next0 = IP4_MAPT_TCP_UDP_NEXT_IP6_LOOKUP; - next1 = IP4_MAPT_TCP_UDP_NEXT_IP6_LOOKUP; - p0 = vlib_get_buffer (vm, pi0); - p1 = vlib_get_buffer (vm, pi1); - - //Accessing pseudo header - pheader0 = vlib_buffer_get_current (p0); - pheader1 = vlib_buffer_get_current (p1); - vlib_buffer_advance (p0, sizeof (*pheader0)); - vlib_buffer_advance (p1, sizeof (*pheader1)); - - if (ip4_to_ip6_tcp_udp (p0, ip4_to_ip6_set_cb, pheader0)) - { - p0->error = error_node->errors[MAP_ERROR_UNKNOWN]; - next0 = IP4_MAPT_TCP_UDP_NEXT_DROP; - } - else - { - if (vnet_buffer (p0)->map_t.mtu < p0->current_length) - { - //Send to fragmentation node if necessary - vnet_buffer (p0)->ip_frag.header_offset = 0; - vnet_buffer (p0)->ip_frag.mtu = vnet_buffer (p0)->map_t.mtu; - vnet_buffer (p0)->ip_frag.next_index = - IP6_FRAG_NEXT_IP6_LOOKUP; - next0 = IP4_MAPT_TCP_UDP_NEXT_IP6_FRAG; - } - } - - if (ip4_to_ip6_tcp_udp (p1, ip4_to_ip6_set_cb, pheader1)) - { - p1->error = error_node->errors[MAP_ERROR_UNKNOWN]; - next1 = IP4_MAPT_TCP_UDP_NEXT_DROP; - } - else - { - if (vnet_buffer (p1)->map_t.mtu < p1->current_length) - { - //Send to fragmentation node if necessary - vnet_buffer (p1)->ip_frag.header_offset = 0; - vnet_buffer (p1)->ip_frag.mtu = vnet_buffer (p1)->map_t.mtu; - vnet_buffer (p1)->ip_frag.next_index = - IP6_FRAG_NEXT_IP6_LOOKUP; - next1 = IP4_MAPT_TCP_UDP_NEXT_IP6_FRAG; - } - } - - vlib_validate_buffer_enqueue_x2 (vm, node, next_index, - to_next, n_left_to_next, pi0, pi1, - next0, next1); - } -#endif - while (n_left_from > 0 && n_left_to_next > 0) { u32 pi0; @@ -443,7 +476,7 @@ ip4_map_t_tcp_udp (vlib_main_t * vm, pheader0 = vlib_buffer_get_current (p0); vlib_buffer_advance (p0, sizeof (*pheader0)); - if (ip4_to_ip6_tcp_udp (p0, ip4_to_ip6_set_cb, pheader0)) + if (map_ip4_to_ip6_tcp_udp (p0, pheader0)) { p0->error = error_node->errors[MAP_ERROR_UNKNOWN]; next0 = IP4_MAPT_TCP_UDP_NEXT_DROP; @@ -453,7 +486,6 @@ ip4_map_t_tcp_udp (vlib_main_t * vm, if (vnet_buffer (p0)->map_t.mtu < p0->current_length) { //Send to fragmentation node if necessary - vnet_buffer (p0)->ip_frag.header_offset = 0; vnet_buffer (p0)->ip_frag.mtu = vnet_buffer (p0)->map_t.mtu; vnet_buffer (p0)->ip_frag.next_index = IP6_FRAG_NEXT_IP6_LOOKUP; @@ -475,14 +507,6 @@ ip4_map_t_classify (vlib_buffer_t * p0, map_domain_t * d0, ip4_header_t * ip40, u16 ip4_len0, i32 * dst_port0, u8 * error0, ip4_mapt_next_t * next0) { - map_main_t *mm = &map_main; - u32 port_offset; - - if (mm->is_ce) - port_offset = 0; - else - port_offset = 2; - if (PREDICT_FALSE (ip4_get_fragment_offset (ip40))) { *next0 = IP4_MAPT_NEXT_MAPT_FRAGMENTED; @@ -501,16 +525,14 @@ ip4_map_t_classify (vlib_buffer_t * p0, map_domain_t * d0, vnet_buffer (p0)->map_t.checksum_offset = 36; *next0 = IP4_MAPT_NEXT_MAPT_TCP_UDP; *error0 = ip4_len0 < 40 ? MAP_ERROR_MALFORMED : *error0; - *dst_port0 = - (i32) * ((u16 *) u8_ptr_add (ip40, sizeof (*ip40) + port_offset)); + *dst_port0 = (i32) * ((u16 *) u8_ptr_add (ip40, sizeof (*ip40) + 2)); } else if (PREDICT_TRUE (ip40->protocol == IP_PROTOCOL_UDP)) { vnet_buffer (p0)->map_t.checksum_offset = 26; *next0 = IP4_MAPT_NEXT_MAPT_TCP_UDP; *error0 = ip4_len0 < 28 ? MAP_ERROR_MALFORMED : *error0; - *dst_port0 = - (i32) * ((u16 *) u8_ptr_add (ip40, sizeof (*ip40) + port_offset)); + *dst_port0 = (i32) * ((u16 *) u8_ptr_add (ip40, sizeof (*ip40) + 2)); } else if (ip40->protocol == IP_PROTOCOL_ICMP) { @@ -539,179 +561,23 @@ ip4_map_t (vlib_main_t * vm, vlib_node_runtime_t * node, vlib_frame_t * frame) from = vlib_frame_vector_args (frame); n_left_from = frame->n_vectors; next_index = node->cached_next_index; - map_main_t *mm = &map_main; vlib_combined_counter_main_t *cm = map_main.domain_counters; - u32 thread_index = vlib_get_thread_index (); + u32 thread_index = vm->thread_index; while (n_left_from > 0) { vlib_get_next_frame (vm, node, next_index, to_next, n_left_to_next); -#ifdef IP4_MAP_T_DUAL_LOOP - while (n_left_from >= 4 && n_left_to_next >= 2) - { - u32 pi0, pi1; - vlib_buffer_t *p0, *p1; - ip4_header_t *ip40, *ip41; - map_domain_t *d0, *d1; - ip4_mapt_next_t next0 = 0, next1 = 0; - u16 ip4_len0, ip4_len1; - u8 error0, error1; - i32 map_port0, map_port1; - ip4_mapt_pseudo_header_t *pheader0, *pheader1; - - pi0 = to_next[0] = from[0]; - pi1 = to_next[1] = from[1]; - from += 2; - n_left_from -= 2; - to_next += 2; - n_left_to_next -= 2; - error0 = MAP_ERROR_NONE; - error1 = MAP_ERROR_NONE; - - p0 = vlib_get_buffer (vm, pi0); - p1 = vlib_get_buffer (vm, pi1); - ip40 = vlib_buffer_get_current (p0); - ip41 = vlib_buffer_get_current (p1); - ip4_len0 = clib_host_to_net_u16 (ip40->length); - ip4_len1 = clib_host_to_net_u16 (ip41->length); - - if (PREDICT_FALSE (p0->current_length < ip4_len0 || - ip40->ip_version_and_header_length != 0x45)) - { - error0 = MAP_ERROR_UNKNOWN; - next0 = IP4_MAPT_NEXT_DROP; - } - - if (PREDICT_FALSE (p1->current_length < ip4_len1 || - ip41->ip_version_and_header_length != 0x45)) - { - error1 = MAP_ERROR_UNKNOWN; - next1 = IP4_MAPT_NEXT_DROP; - } - - vnet_buffer (p0)->map_t.map_domain_index = - vnet_buffer (p0)->ip.adj_index[VLIB_TX]; - d0 = ip4_map_get_domain (vnet_buffer (p0)->map_t.map_domain_index); - vnet_buffer (p1)->map_t.map_domain_index = - vnet_buffer (p1)->ip.adj_index[VLIB_TX]; - d1 = ip4_map_get_domain (vnet_buffer (p1)->map_t.map_domain_index); - - vnet_buffer (p0)->map_t.mtu = d0->mtu ? d0->mtu : ~0; - vnet_buffer (p1)->map_t.mtu = d1->mtu ? d1->mtu : ~0; - - map_port0 = -1; - map_port1 = -1; - - ip4_map_t_classify (p0, d0, ip40, ip4_len0, &map_port0, &error0, - &next0); - ip4_map_t_classify (p1, d1, ip41, ip4_len1, &map_port1, &error1, - &next1); - - //Add MAP-T pseudo header in front of the packet - vlib_buffer_advance (p0, -sizeof (*pheader0)); - vlib_buffer_advance (p1, -sizeof (*pheader1)); - pheader0 = vlib_buffer_get_current (p0); - pheader1 = vlib_buffer_get_current (p1); - - //Save addresses within the packet - if (mm->is_ce) - { - ip4_map_t_embedded_address (d0, &pheader0->daddr, - &ip40->dst_address); - ip4_map_t_embedded_address (d1, &pheader1->daddr, - &ip41->dst_address); - pheader0->saddr.as_u64[0] = - map_get_pfx_net (d0, ip40->src_address.as_u32, - (u16) map_port0); - pheader0->saddr.as_u64[1] = - map_get_sfx_net (d0, ip40->src_address.as_u32, - (u16) map_port0); - pheader1->saddr.as_u64[0] = - map_get_pfx_net (d1, ip41->src_address.as_u32, - (u16) map_port1); - pheader1->saddr.as_u64[1] = - map_get_sfx_net (d1, ip41->src_address.as_u32, - (u16) map_port1); - } - else - { - ip4_map_t_embedded_address (d0, &pheader0->saddr, - &ip40->src_address); - ip4_map_t_embedded_address (d1, &pheader1->saddr, - &ip41->src_address); - pheader0->daddr.as_u64[0] = - map_get_pfx_net (d0, ip40->dst_address.as_u32, - (u16) map_port0); - pheader0->daddr.as_u64[1] = - map_get_sfx_net (d0, ip40->dst_address.as_u32, - (u16) map_port0); - pheader1->daddr.as_u64[0] = - map_get_pfx_net (d1, ip41->dst_address.as_u32, - (u16) map_port1); - pheader1->daddr.as_u64[1] = - map_get_sfx_net (d1, ip41->dst_address.as_u32, - (u16) map_port1); - } - - if (PREDICT_FALSE - (ip4_is_first_fragment (ip40) && (map_port0 != -1) - && (d0->ea_bits_len != 0 || !d0->rules) - && ip4_map_fragment_cache (ip40, map_port0))) - { - error0 = MAP_ERROR_FRAGMENT_MEMORY; - } - - if (PREDICT_FALSE - (ip4_is_first_fragment (ip41) && (map_port1 != -1) - && (d1->ea_bits_len != 0 || !d1->rules) - && ip4_map_fragment_cache (ip41, map_port1))) - { - error1 = MAP_ERROR_FRAGMENT_MEMORY; - } - - if (PREDICT_TRUE - (error0 == MAP_ERROR_NONE && next0 != IP4_MAPT_NEXT_MAPT_ICMP)) - { - vlib_increment_combined_counter (cm + MAP_DOMAIN_COUNTER_TX, - thread_index, - vnet_buffer (p0)-> - map_t.map_domain_index, 1, - clib_net_to_host_u16 - (ip40->length)); - } - - if (PREDICT_TRUE - (error1 == MAP_ERROR_NONE && next1 != IP4_MAPT_NEXT_MAPT_ICMP)) - { - vlib_increment_combined_counter (cm + MAP_DOMAIN_COUNTER_TX, - thread_index, - vnet_buffer (p1)-> - map_t.map_domain_index, 1, - clib_net_to_host_u16 - (ip41->length)); - } - - next0 = (error0 != MAP_ERROR_NONE) ? IP4_MAPT_NEXT_DROP : next0; - next1 = (error1 != MAP_ERROR_NONE) ? IP4_MAPT_NEXT_DROP : next1; - p0->error = error_node->errors[error0]; - p1->error = error_node->errors[error1]; - vlib_validate_buffer_enqueue_x2 (vm, node, next_index, to_next, - n_left_to_next, pi0, pi1, next0, - next1); - } -#endif - while (n_left_from > 0 && n_left_to_next > 0) { u32 pi0; vlib_buffer_t *p0; ip4_header_t *ip40; map_domain_t *d0; - ip4_mapt_next_t next0; + ip4_mapt_next_t next0 = 0; u16 ip4_len0; u8 error0; - i32 map_port0; + i32 dst_port0; ip4_mapt_pseudo_header_t *pheader0; pi0 = to_next[0] = from[0]; @@ -728,54 +594,52 @@ ip4_map_t (vlib_main_t * vm, vlib_node_runtime_t * node, vlib_frame_t * frame) ip40->ip_version_and_header_length != 0x45)) { error0 = MAP_ERROR_UNKNOWN; - next0 = IP4_MAPT_NEXT_DROP; } - vnet_buffer (p0)->map_t.map_domain_index = - vnet_buffer (p0)->ip.adj_index[VLIB_TX]; - d0 = ip4_map_get_domain (vnet_buffer (p0)->map_t.map_domain_index); + d0 = ip4_map_get_domain (&ip40->dst_address, + &vnet_buffer (p0)->map_t.map_domain_index, + &error0); + + if (!d0) + { /* Guess it wasn't for us */ + vnet_feature_next (&next0, p0); + goto exit; + } vnet_buffer (p0)->map_t.mtu = d0->mtu ? d0->mtu : ~0; - map_port0 = -1; - ip4_map_t_classify (p0, d0, ip40, ip4_len0, &map_port0, &error0, + dst_port0 = -1; + ip4_map_t_classify (p0, d0, ip40, ip4_len0, &dst_port0, &error0, &next0); + /* Verify that port is not among the well-known ports */ + if ((d0->psid_length > 0 && d0->psid_offset > 0) + && (clib_net_to_host_u16 (dst_port0) < + (0x1 << (16 - d0->psid_offset)))) + { + error0 = MAP_ERROR_SEC_CHECK; + } + //Add MAP-T pseudo header in front of the packet vlib_buffer_advance (p0, -sizeof (*pheader0)); pheader0 = vlib_buffer_get_current (p0); //Save addresses within the packet - if (mm->is_ce) - { - ip4_map_t_embedded_address (d0, &pheader0->daddr, - &ip40->dst_address); - pheader0->saddr.as_u64[0] = - map_get_pfx_net (d0, ip40->src_address.as_u32, - (u16) map_port0); - pheader0->saddr.as_u64[1] = - map_get_sfx_net (d0, ip40->src_address.as_u32, - (u16) map_port0); - } - else - { - ip4_map_t_embedded_address (d0, &pheader0->saddr, - &ip40->src_address); - pheader0->daddr.as_u64[0] = - map_get_pfx_net (d0, ip40->dst_address.as_u32, - (u16) map_port0); - pheader0->daddr.as_u64[1] = - map_get_sfx_net (d0, ip40->dst_address.as_u32, - (u16) map_port0); - } - - //It is important to cache at this stage because the result might be necessary - //for packets within the same vector. - //Actually, this approach even provides some limited out-of-order fragments support + ip4_map_t_embedded_address (d0, &pheader0->saddr, + &ip40->src_address); + pheader0->daddr.as_u64[0] = + map_get_pfx_net (d0, ip40->dst_address.as_u32, (u16) dst_port0); + pheader0->daddr.as_u64[1] = + map_get_sfx_net (d0, ip40->dst_address.as_u32, (u16) dst_port0); + + // It is important to cache at this stage because the result + // might be necessary for packets within the same vector. + // Actually, this approach even provides some limited + // out-of-order fragments support if (PREDICT_FALSE - (ip4_is_first_fragment (ip40) && (map_port0 != -1) + (ip4_is_first_fragment (ip40) && (dst_port0 != -1) && (d0->ea_bits_len != 0 || !d0->rules) - && ip4_map_fragment_cache (ip40, map_port0))) + && ip4_map_fragment_cache (ip40, dst_port0))) { error0 = MAP_ERROR_UNKNOWN; } @@ -793,6 +657,7 @@ ip4_map_t (vlib_main_t * vm, vlib_node_runtime_t * node, vlib_frame_t * frame) next0 = (error0 != MAP_ERROR_NONE) ? IP4_MAPT_NEXT_DROP : next0; p0->error = error_node->errors[error0]; + exit: vlib_validate_buffer_enqueue_x1 (vm, node, next_index, to_next, n_left_to_next, pi0, next0); @@ -808,6 +673,11 @@ static char *map_t_error_strings[] = { #undef _ }; +VNET_FEATURE_INIT (ip4_map_t_feature, static) = +{ +.arc_name = "ip4-unicast",.node_name = "ip4-map-t",.runs_before = + VNET_FEATURES ("ip4-flow-classify"),}; + /* *INDENT-OFF* */ VLIB_REGISTER_NODE(ip4_map_t_fragmented_node) = { .function = ip4_map_t_fragmented,