2 * Copyright (c) 2018 Cisco and/or its affiliates.
3 * Licensed under the Apache License, Version 2.0 (the "License");
4 * you may not use this file except in compliance with the License.
5 * You may obtain a copy of the License at:
7 * http://www.apache.org/licenses/LICENSE-2.0
9 * Unless required by applicable law or agreed to in writing, software
10 * distributed under the License is distributed on an "AS IS" BASIS,
11 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
12 * See the License for the specific language governing permissions and
13 * limitations under the License.
16 * @brief The NAT inline functions
19 #ifndef __included_nat_inlines_h__
20 #define __included_nat_inlines_h__
22 #include <vnet/fib/ip4_fib.h>
24 #include <nat/nat_ha.h>
27 nat_pre_node_fn_inline (vlib_main_t * vm,
28 vlib_node_runtime_t * node,
29 vlib_frame_t * frame, u32 def_next)
31 u32 n_left_from, *from, *to_next;
34 from = vlib_frame_vector_args (frame);
35 n_left_from = frame->n_vectors;
36 next_index = node->cached_next_index;
38 while (n_left_from > 0)
42 vlib_get_next_frame (vm, node, next_index, to_next, n_left_to_next);
44 while (n_left_from >= 4 && n_left_to_next >= 2)
47 u32 arc_next0, arc_next1;
49 vlib_buffer_t *b0, *b1;
51 /* Prefetch next iteration. */
53 vlib_buffer_t *p2, *p3;
55 p2 = vlib_get_buffer (vm, from[2]);
56 p3 = vlib_get_buffer (vm, from[3]);
58 vlib_prefetch_buffer_header (p2, LOAD);
59 vlib_prefetch_buffer_header (p3, LOAD);
61 CLIB_PREFETCH (p2->data, CLIB_CACHE_LINE_BYTES, STORE);
62 CLIB_PREFETCH (p3->data, CLIB_CACHE_LINE_BYTES, STORE);
65 /* speculatively enqueue b0 and b1 to the current next frame */
66 to_next[0] = bi0 = from[0];
67 to_next[1] = bi1 = from[1];
73 b0 = vlib_get_buffer (vm, bi0);
74 b1 = vlib_get_buffer (vm, bi1);
79 vnet_feature_next (&arc_next0, b0);
80 vnet_feature_next (&arc_next1, b1);
82 nat_buffer_opaque (b0)->arc_next = arc_next0;
83 nat_buffer_opaque (b1)->arc_next = arc_next1;
85 if (PREDICT_FALSE ((node->flags & VLIB_NODE_FLAG_TRACE)))
87 if (b0->flags & VLIB_BUFFER_IS_TRACED)
90 vlib_add_trace (vm, node, b0, sizeof (*t));
91 t->next_index = next0;
93 if (b1->flags & VLIB_BUFFER_IS_TRACED)
96 vlib_add_trace (vm, node, b0, sizeof (*t));
97 t->next_index = next0;
101 /* verify speculative enqueues, maybe switch current next frame */
102 vlib_validate_buffer_enqueue_x2 (vm, node, next_index,
103 to_next, n_left_to_next,
104 bi0, bi1, next0, next1);
107 while (n_left_from > 0 && n_left_to_next > 0)
114 /* speculatively enqueue b0 to the current next frame */
122 b0 = vlib_get_buffer (vm, bi0);
124 vnet_feature_next (&arc_next0, b0);
125 nat_buffer_opaque (b0)->arc_next = arc_next0;
127 if (PREDICT_FALSE ((node->flags & VLIB_NODE_FLAG_TRACE)
128 && (b0->flags & VLIB_BUFFER_IS_TRACED)))
130 nat_pre_trace_t *t = vlib_add_trace (vm, node, b0, sizeof (*t));
131 t->next_index = next0;
134 /* verify speculative enqueue, maybe switch current next frame */
135 vlib_validate_buffer_enqueue_x1 (vm, node, next_index,
136 to_next, n_left_to_next,
140 vlib_put_next_frame (vm, node, next_index, n_left_to_next);
143 return frame->n_vectors;
147 ip_proto_to_snat_proto (u8 ip_proto)
151 snat_proto = (ip_proto == IP_PROTOCOL_UDP) ? SNAT_PROTOCOL_UDP : snat_proto;
152 snat_proto = (ip_proto == IP_PROTOCOL_TCP) ? SNAT_PROTOCOL_TCP : snat_proto;
154 (ip_proto == IP_PROTOCOL_ICMP) ? SNAT_PROTOCOL_ICMP : snat_proto;
156 (ip_proto == IP_PROTOCOL_ICMP6) ? SNAT_PROTOCOL_ICMP : snat_proto;
162 snat_proto_to_ip_proto (snat_protocol_t snat_proto)
166 ip_proto = (snat_proto == SNAT_PROTOCOL_UDP) ? IP_PROTOCOL_UDP : ip_proto;
167 ip_proto = (snat_proto == SNAT_PROTOCOL_TCP) ? IP_PROTOCOL_TCP : ip_proto;
168 ip_proto = (snat_proto == SNAT_PROTOCOL_ICMP) ? IP_PROTOCOL_ICMP : ip_proto;
173 static_always_inline u8
174 icmp_is_error_message (icmp46_header_t * icmp)
178 case ICMP4_destination_unreachable:
179 case ICMP4_time_exceeded:
180 case ICMP4_parameter_problem:
181 case ICMP4_source_quench:
183 case ICMP4_alternate_host_address:
190 is_interface_addr (snat_main_t * sm, vlib_node_runtime_t * node,
191 u32 sw_if_index0, u32 ip4_addr)
193 snat_runtime_t *rt = (snat_runtime_t *) node->runtime_data;
194 ip4_address_t *first_int_addr;
196 if (PREDICT_FALSE (rt->cached_sw_if_index != sw_if_index0))
199 ip4_interface_first_address (sm->ip4_main, sw_if_index0,
200 0 /* just want the address */ );
201 rt->cached_sw_if_index = sw_if_index0;
203 rt->cached_ip4_address = first_int_addr->as_u32;
205 rt->cached_ip4_address = 0;
208 if (PREDICT_FALSE (ip4_addr == rt->cached_ip4_address))
215 maximum_sessions_exceeded (snat_main_t * sm, u32 thread_index)
217 if (pool_elts (sm->per_thread_data[thread_index].sessions) >=
218 sm->max_translations)
225 nat_send_all_to_node (vlib_main_t * vm, u32 * bi_vector,
226 vlib_node_runtime_t * node, vlib_error_t * error,
229 u32 n_left_from, *from, next_index, *to_next, n_left_to_next;
232 n_left_from = vec_len (bi_vector);
233 next_index = node->cached_next_index;
234 while (n_left_from > 0)
236 vlib_get_next_frame (vm, node, next_index, to_next, n_left_to_next);
237 while (n_left_from > 0 && n_left_to_next > 0)
239 u32 bi0 = to_next[0] = from[0];
244 vlib_buffer_t *p0 = vlib_get_buffer (vm, bi0);
247 vlib_validate_buffer_enqueue_x1 (vm, node, next_index, to_next,
248 n_left_to_next, bi0, next);
250 vlib_put_next_frame (vm, node, next_index, n_left_to_next);
255 user_session_increment (snat_main_t * sm, snat_user_t * u, u8 is_static)
257 if (u->nsessions + u->nstaticsessions < sm->max_translations_per_user)
260 u->nstaticsessions++;
267 nat44_delete_user_with_no_session (snat_main_t * sm, snat_user_t * u,
270 clib_bihash_kv_8_8_t kv;
271 snat_user_key_t u_key;
272 snat_main_per_thread_data_t *tsm = vec_elt_at_index (sm->per_thread_data,
275 if (u->nstaticsessions == 0 && u->nsessions == 0)
277 u_key.addr.as_u32 = u->addr.as_u32;
278 u_key.fib_index = u->fib_index;
279 kv.key = u_key.as_u64;
280 pool_put_index (tsm->list_pool, u->sessions_per_user_list_head_index);
281 pool_put (tsm->users, u);
282 clib_bihash_add_del_8_8 (&tsm->user_hash, &kv, 0);
283 vlib_set_simple_counter (&sm->total_users, thread_index, 0,
284 pool_elts (tsm->users));
289 nat44_delete_session (snat_main_t * sm, snat_session_t * ses,
292 snat_main_per_thread_data_t *tsm = vec_elt_at_index (sm->per_thread_data,
294 clib_bihash_kv_8_8_t kv, value;
296 const snat_user_key_t u_key = {
297 .addr = ses->in2out.addr,
298 .fib_index = ses->in2out.fib_index
300 const u8 u_static = snat_is_session_static (ses);
302 clib_dlist_remove (tsm->list_pool, ses->per_user_index);
303 pool_put_index (tsm->list_pool, ses->per_user_index);
304 pool_put (tsm->sessions, ses);
305 vlib_set_simple_counter (&sm->total_sessions, thread_index, 0,
306 pool_elts (tsm->sessions));
308 kv.key = u_key.as_u64;
309 if (!clib_bihash_search_8_8 (&tsm->user_hash, &kv, &value))
311 u = pool_elt_at_index (tsm->users, value.value);
313 u->nstaticsessions--;
317 nat44_delete_user_with_no_session (sm, u, thread_index);
321 /** \brief Set TCP session state.
322 @return 1 if session was closed, otherwise 0
325 nat44_set_tcp_session_state_i2o (snat_main_t * sm, snat_session_t * ses,
326 tcp_header_t * tcp, u32 thread_index)
328 if ((ses->state == 0) && (tcp->flags & TCP_FLAG_RST))
329 ses->state = NAT44_SES_RST;
330 if ((ses->state == NAT44_SES_RST) && !(tcp->flags & TCP_FLAG_RST))
332 if ((tcp->flags & TCP_FLAG_ACK) && (ses->state & NAT44_SES_I2O_SYN) &&
333 (ses->state & NAT44_SES_O2I_SYN))
335 if (tcp->flags & TCP_FLAG_SYN)
336 ses->state |= NAT44_SES_I2O_SYN;
337 if (tcp->flags & TCP_FLAG_FIN)
339 ses->i2o_fin_seq = clib_net_to_host_u32 (tcp->seq_number);
340 ses->state |= NAT44_SES_I2O_FIN;
342 if ((tcp->flags & TCP_FLAG_ACK) && (ses->state & NAT44_SES_O2I_FIN))
344 if (clib_net_to_host_u32 (tcp->ack_number) > ses->o2i_fin_seq)
345 ses->state |= NAT44_SES_O2I_FIN_ACK;
347 if (nat44_is_ses_closed (ses)
348 && !(ses->flags & SNAT_SESSION_FLAG_OUTPUT_FEATURE))
350 nat_free_session_data (sm, ses, thread_index, 0);
351 nat44_delete_session (sm, ses, thread_index);
358 nat44_set_tcp_session_state_o2i (snat_main_t * sm, snat_session_t * ses,
359 tcp_header_t * tcp, u32 thread_index)
361 if ((ses->state == 0) && (tcp->flags & TCP_FLAG_RST))
362 ses->state = NAT44_SES_RST;
363 if ((ses->state == NAT44_SES_RST) && !(tcp->flags & TCP_FLAG_RST))
365 if ((tcp->flags & TCP_FLAG_ACK) && (ses->state & NAT44_SES_I2O_SYN) &&
366 (ses->state & NAT44_SES_O2I_SYN))
368 if (tcp->flags & TCP_FLAG_SYN)
369 ses->state |= NAT44_SES_O2I_SYN;
370 if (tcp->flags & TCP_FLAG_FIN)
372 ses->o2i_fin_seq = clib_net_to_host_u32 (tcp->seq_number);
373 ses->state |= NAT44_SES_O2I_FIN;
375 if ((tcp->flags & TCP_FLAG_ACK) && (ses->state & NAT44_SES_I2O_FIN))
377 if (clib_net_to_host_u32 (tcp->ack_number) > ses->i2o_fin_seq)
378 ses->state |= NAT44_SES_I2O_FIN_ACK;
380 if (nat44_is_ses_closed (ses))
382 nat_free_session_data (sm, ses, thread_index, 0);
383 nat44_delete_session (sm, ses, thread_index);
390 nat44_session_get_timeout (snat_main_t * sm, snat_session_t * s)
392 switch (s->in2out.protocol)
394 case SNAT_PROTOCOL_ICMP:
395 return sm->icmp_timeout;
396 case SNAT_PROTOCOL_UDP:
397 return sm->udp_timeout;
398 case SNAT_PROTOCOL_TCP:
401 return sm->tcp_transitory_timeout;
403 return sm->tcp_established_timeout;
406 return sm->udp_timeout;
413 nat44_session_update_counters (snat_session_t * s, f64 now, uword bytes,
418 s->total_bytes += bytes;
419 nat_ha_sref (&s->out2in.addr, s->out2in.port, &s->ext_host_addr,
420 s->ext_host_port, s->out2in.protocol, s->out2in.fib_index,
421 s->total_pkts, s->total_bytes, thread_index,
422 &s->ha_last_refreshed, now);
425 /** \brief Per-user LRU list maintenance */
427 nat44_session_update_lru (snat_main_t * sm, snat_session_t * s,
430 clib_dlist_remove (sm->per_thread_data[thread_index].list_pool,
432 clib_dlist_addtail (sm->per_thread_data[thread_index].list_pool,
433 s->per_user_list_head_index, s->per_user_index);
437 make_ed_kv (clib_bihash_kv_16_8_t * kv, ip4_address_t * l_addr,
438 ip4_address_t * r_addr, u8 proto, u32 fib_index, u16 l_port,
441 nat_ed_ses_key_t *key = (nat_ed_ses_key_t *) kv->key;
443 key->l_addr.as_u32 = l_addr->as_u32;
444 key->r_addr.as_u32 = r_addr->as_u32;
445 key->fib_index = fib_index;
447 key->l_port = l_port;
448 key->r_port = r_port;
454 make_sm_kv (clib_bihash_kv_8_8_t * kv, ip4_address_t * addr, u8 proto,
455 u32 fib_index, u16 port)
457 snat_session_key_t key;
459 key.addr.as_u32 = addr->as_u32;
461 key.protocol = proto;
462 key.fib_index = fib_index;
464 kv->key = key.as_u64;
468 static_always_inline int
469 get_icmp_i2o_ed_key (ip4_header_t * ip0, nat_ed_ses_key_t * p_key0)
471 icmp46_header_t *icmp0;
472 nat_ed_ses_key_t key0;
473 icmp_echo_header_t *echo0, *inner_echo0 = 0;
474 ip4_header_t *inner_ip0 = 0;
476 icmp46_header_t *inner_icmp0;
478 icmp0 = (icmp46_header_t *) ip4_next_header (ip0);
479 echo0 = (icmp_echo_header_t *) (icmp0 + 1);
481 if (!icmp_is_error_message (icmp0))
483 key0.proto = IP_PROTOCOL_ICMP;
484 key0.l_addr = ip0->src_address;
485 key0.r_addr = ip0->dst_address;
486 key0.l_port = echo0->identifier;
491 inner_ip0 = (ip4_header_t *) (echo0 + 1);
492 l4_header = ip4_next_header (inner_ip0);
493 key0.proto = inner_ip0->protocol;
494 key0.r_addr = inner_ip0->src_address;
495 key0.l_addr = inner_ip0->dst_address;
496 switch (ip_proto_to_snat_proto (inner_ip0->protocol))
498 case SNAT_PROTOCOL_ICMP:
499 inner_icmp0 = (icmp46_header_t *) l4_header;
500 inner_echo0 = (icmp_echo_header_t *) (inner_icmp0 + 1);
502 key0.l_port = inner_echo0->identifier;
504 case SNAT_PROTOCOL_UDP:
505 case SNAT_PROTOCOL_TCP:
506 key0.l_port = ((tcp_udp_header_t *) l4_header)->dst_port;
507 key0.r_port = ((tcp_udp_header_t *) l4_header)->src_port;
510 return NAT_IN2OUT_ED_ERROR_UNSUPPORTED_PROTOCOL;
518 static_always_inline int
519 get_icmp_o2i_ed_key (ip4_header_t * ip0, nat_ed_ses_key_t * p_key0)
521 icmp46_header_t *icmp0;
522 nat_ed_ses_key_t key0;
523 icmp_echo_header_t *echo0, *inner_echo0 = 0;
524 ip4_header_t *inner_ip0;
526 icmp46_header_t *inner_icmp0;
528 icmp0 = (icmp46_header_t *) ip4_next_header (ip0);
529 echo0 = (icmp_echo_header_t *) (icmp0 + 1);
531 if (!icmp_is_error_message (icmp0))
533 key0.proto = IP_PROTOCOL_ICMP;
534 key0.l_addr = ip0->dst_address;
535 key0.r_addr = ip0->src_address;
536 key0.l_port = echo0->identifier;
541 inner_ip0 = (ip4_header_t *) (echo0 + 1);
542 l4_header = ip4_next_header (inner_ip0);
543 key0.proto = inner_ip0->protocol;
544 key0.l_addr = inner_ip0->src_address;
545 key0.r_addr = inner_ip0->dst_address;
546 switch (ip_proto_to_snat_proto (inner_ip0->protocol))
548 case SNAT_PROTOCOL_ICMP:
549 inner_icmp0 = (icmp46_header_t *) l4_header;
550 inner_echo0 = (icmp_echo_header_t *) (inner_icmp0 + 1);
551 key0.l_port = inner_echo0->identifier;
554 case SNAT_PROTOCOL_UDP:
555 case SNAT_PROTOCOL_TCP:
556 key0.l_port = ((tcp_udp_header_t *) l4_header)->src_port;
557 key0.r_port = ((tcp_udp_header_t *) l4_header)->dst_port;
568 mss_clamping (snat_main_t * sm, tcp_header_t * tcp, ip_csum_t * sum)
571 u8 opt_len, opts_len, kind;
574 if (!(sm->mss_clamping && tcp_syn (tcp)))
577 opts_len = (tcp_doff (tcp) << 2) - sizeof (tcp_header_t);
578 data = (u8 *) (tcp + 1);
579 for (; opts_len > 0; opts_len -= opt_len, data += opt_len)
583 if (kind == TCP_OPTION_EOL)
585 else if (kind == TCP_OPTION_NOOP)
596 if (opt_len < 2 || opt_len > opts_len)
600 if (kind == TCP_OPTION_MSS)
602 mss = *(u16 *) (data + 2);
603 if (clib_net_to_host_u16 (mss) > sm->mss_clamping)
606 ip_csum_update (*sum, mss, sm->mss_value_net, ip4_header_t,
608 clib_memcpy_fast (data + 2, &sm->mss_value_net, 2);
616 * @brief Check if packet should be translated
618 * Packets aimed at outside interface and external address with active session
619 * should be translated.
622 * @param rt NAT runtime data
623 * @param sw_if_index0 index of the inside interface
624 * @param ip0 IPv4 header
625 * @param proto0 NAT protocol
626 * @param rx_fib_index0 RX FIB index
628 * @returns 0 if packet should be translated otherwise 1
631 snat_not_translate_fast (snat_main_t * sm, vlib_node_runtime_t * node,
632 u32 sw_if_index0, ip4_header_t * ip0, u32 proto0,
638 fib_node_index_t fei = FIB_NODE_INDEX_INVALID;
639 nat_outside_fib_t *outside_fib;
641 .fp_proto = FIB_PROTOCOL_IP4,
644 .ip4.as_u32 = ip0->dst_address.as_u32,
649 /* Don't NAT packet aimed at the intfc address */
650 if (PREDICT_FALSE (is_interface_addr (sm, node, sw_if_index0,
651 ip0->dst_address.as_u32)))
654 fei = fib_table_lookup (rx_fib_index0, &pfx);
655 if (FIB_NODE_INDEX_INVALID != fei)
657 u32 sw_if_index = fib_entry_get_resolving_interface (fei);
658 if (sw_if_index == ~0)
660 vec_foreach (outside_fib, sm->outside_fibs)
662 fei = fib_table_lookup (outside_fib->fib_index, &pfx);
663 if (FIB_NODE_INDEX_INVALID != fei)
665 sw_if_index = fib_entry_get_resolving_interface (fei);
666 if (sw_if_index != ~0)
671 if (sw_if_index == ~0)
676 pool_foreach (i, sm->interfaces, ({
677 /* NAT packet aimed at outside interface */
678 if ((nat_interface_is_outside (i)) && (sw_if_index == i->sw_if_index))
687 #endif /* __included_nat_inlines_h__ */
690 * fd.io coding-style-patch-verification: ON
693 * eval: (c-set-style "gnu")