nat: twice-nat static mapping pool address
[vpp.git] / src / plugins / nat / in2out_ed.c
index e5f29e4..448e967 100644 (file)
 #include <nat/nat44/ed_inlines.h>
 #include <nat/lib/nat_inlines.h>
 
+/* number of attempts to get a port for ED overloading algorithm, if rolling
+ * a dice this many times doesn't produce a free port, it's treated
+ * as if there were no free ports available to conserve resources */
+#define ED_PORT_ALLOC_ATTEMPTS (10)
+
 static char *nat_in2out_ed_error_strings[] = {
 #define _(sym,string) string,
   foreach_nat_in2out_ed_error
@@ -217,17 +222,17 @@ nat_ed_alloc_addr_and_port (snat_main_t * sm, u32 rx_fib_index,
     if (a->fib_index == rx_fib_index)                                        \
       {                                                                      \
         /* first try port suggested by caller */                             \
-        u16 port = clib_net_to_host_u16 (*outside_port);                   \
+        u16 port = clib_net_to_host_u16 (*outside_port);                     \
         u16 port_offset = port - port_thread_offset;                         \
         if (port <= port_thread_offset ||                                    \
             port > port_thread_offset + port_per_thread)                     \
           {                                                                  \
             /* need to pick a different port, suggested port doesn't fit in  \
              * this thread's port range */                                   \
-            port_offset = snat_random_port (1, port_per_thread);             \
+            port_offset = snat_random_port (0, port_per_thread - 1);         \
             port = port_thread_offset + port_offset;                         \
           }                                                                  \
-        u16 attempts = port_per_thread;                                      \
+        u16 attempts = ED_PORT_ALLOC_ATTEMPTS;                               \
         do                                                                   \
           {                                                                  \
             init_ed_kv (out2in_ed_kv, a->addr, clib_host_to_net_u16 (port),  \
@@ -240,11 +245,11 @@ nat_ed_alloc_addr_and_port (snat_main_t * sm, u32 rx_fib_index,
                 ++a->busy_##n##_port_refcounts[port];                        \
                 a->busy_##n##_ports_per_thread[thread_index]++;              \
                 a->busy_##n##_ports++;                                       \
-                *outside_addr = a->addr;                                   \
-                *outside_port = clib_host_to_net_u16 (port);               \
+                *outside_addr = a->addr;                                     \
+                *outside_port = clib_host_to_net_u16 (port);                 \
                 return 0;                                                    \
               }                                                              \
-            port_offset = (port_offset + 1) % port_per_thread;               \
+            port_offset = snat_random_port (0, port_per_thread - 1);         \
             port = port_thread_offset + port_offset;                         \
             --attempts;                                                      \
           }                                                                  \
@@ -365,7 +370,7 @@ slow_path_ed (snat_main_t * sm,
   /* First try to match static mapping by local address and port */
   if (snat_static_mapping_match
       (sm, l_addr, l_port, rx_fib_index, nat_proto, &sm_addr, &sm_port,
-       &sm_fib_index, 0, 0, 0, &lb, 0, &identity_nat))
+       &sm_fib_index, 0, 0, 0, &lb, 0, &identity_nat, 0))
     {
       s = nat_ed_session_alloc (sm, thread_index, now, proto);
       ASSERT (s);
@@ -483,6 +488,8 @@ slow_path_ed (snat_main_t * sm,
               &s->ext_host_nat_addr, s->ext_host_nat_port,
               s->nat_proto, s->in2out.fib_index, s->flags, thread_index, 0);
 
+  per_vrf_sessions_register_session (s, thread_index);
+
   return next;
 }
 
@@ -501,12 +508,13 @@ nat44_ed_not_translate (snat_main_t * sm, vlib_node_runtime_t * node,
   if (clib_bihash_search_16_8 (&sm->out2in_ed, &kv, &value))
     {
       /* or is static mappings */
-      ip4_address_t dummy_addr;
-      u16 dummy_port;
-      u32 dummy_fib_index;
+      ip4_address_t placeholder_addr;
+      u16 placeholder_port;
+      u32 placeholder_fib_index;
       if (!snat_static_mapping_match
          (sm, ip->dst_address, udp->dst_port, sm->outside_fib_index, proto,
-          &dummy_addr, &dummy_port, &dummy_fib_index, 1, 0, 0, 0, 0, 0))
+          &placeholder_addr, &placeholder_port, &placeholder_fib_index, 1, 0,
+          0, 0, 0, 0, 0))
        return 0;
     }
   else
@@ -880,6 +888,8 @@ nat44_ed_in2out_unknown_proto (snat_main_t * sm,
                  s - tsm->sessions);
       if (clib_bihash_add_del_16_8 (&sm->out2in_ed, &s_kv, 1))
        nat_elog_notice ("out2in key add failed");
+
+      per_vrf_sessions_register_session (s, thread_index);
     }
 
   /* Update IP checksum */
@@ -909,206 +919,179 @@ nat44_ed_in2out_fast_path_node_fn_inline (vlib_main_t * vm,
                                          vlib_frame_t * frame,
                                          int is_output_feature)
 {
-  u32 n_left_from, *from, *to_next, pkts_processed = 0, stats_node_index;
-  nat_next_t next_index;
+  u32 n_left_from, *from;
   snat_main_t *sm = &snat_main;
   f64 now = vlib_time_now (vm);
   u32 thread_index = vm->thread_index;
   snat_main_per_thread_data_t *tsm = &sm->per_thread_data[thread_index];
-  u32 tcp_packets = 0, udp_packets = 0, icmp_packets = 0, other_packets =
-    0, def_slow;
-
-  def_slow = is_output_feature ? NAT_NEXT_IN2OUT_ED_OUTPUT_SLOW_PATH :
-    NAT_NEXT_IN2OUT_ED_SLOW_PATH;
-
-  stats_node_index = sm->ed_in2out_node_index;
+  u32 def_slow = is_output_feature ? NAT_NEXT_IN2OUT_ED_OUTPUT_SLOW_PATH
+    : NAT_NEXT_IN2OUT_ED_SLOW_PATH;
 
   from = vlib_frame_vector_args (frame);
   n_left_from = frame->n_vectors;
-  next_index = node->cached_next_index;
+
+  vlib_buffer_t *bufs[VLIB_FRAME_SIZE], **b = bufs;
+  u16 nexts[VLIB_FRAME_SIZE], *next = nexts;
+  vlib_get_buffers (vm, from, b, n_left_from);
 
   while (n_left_from > 0)
     {
-      u32 n_left_to_next;
+      vlib_buffer_t *b0;
+      u32 sw_if_index0, rx_fib_index0, proto0, new_addr0, old_addr0,
+       iph_offset0 = 0;
+      u16 old_port0, new_port0;
+      ip4_header_t *ip0;
+      udp_header_t *udp0;
+      tcp_header_t *tcp0;
+      snat_session_t *s0 = 0;
+      clib_bihash_kv_16_8_t kv0, value0;
+      ip_csum_t sum0;
+
+      b0 = *b;
+      b++;
+
+      /* Prefetch next iteration. */
+      if (PREDICT_TRUE (n_left_from >= 2))
+       {
+         vlib_buffer_t *p2;
 
-      vlib_get_next_frame (vm, node, next_index, to_next, n_left_to_next);
+         p2 = *b;
 
-      while (n_left_from > 0 && n_left_to_next > 0)
+         vlib_prefetch_buffer_header (p2, LOAD);
+
+         CLIB_PREFETCH (p2->data, CLIB_CACHE_LINE_BYTES, LOAD);
+       }
+
+      if (is_output_feature)
        {
-         u32 bi0;
-         vlib_buffer_t *b0;
-         u32 next0, sw_if_index0, rx_fib_index0, iph_offset0 = 0, proto0,
-           new_addr0, old_addr0;
-         u16 old_port0, new_port0;
-         ip4_header_t *ip0;
-         udp_header_t *udp0;
-         tcp_header_t *tcp0;
-         snat_session_t *s0 = 0;
-         clib_bihash_kv_16_8_t kv0, value0;
-         ip_csum_t sum0;
-
-         /* speculatively enqueue b0 to the current next frame */
-         bi0 = from[0];
-         to_next[0] = bi0;
-         from += 1;
-         to_next += 1;
-         n_left_from -= 1;
-         n_left_to_next -= 1;
-
-         b0 = vlib_get_buffer (vm, bi0);
+         vnet_feature_next (&vnet_buffer2 (b0)->nat.arc_next, b0);
+         iph_offset0 = vnet_buffer (b0)->ip.reass.save_rewrite_length;
+       }
 
-         if (is_output_feature)
-           {
-             vnet_feature_next (&vnet_buffer2 (b0)->nat.arc_next, b0);
-             iph_offset0 = vnet_buffer (b0)->ip.reass.save_rewrite_length;
-           }
+      next[0] = vnet_buffer2 (b0)->nat.arc_next;
 
-         next0 = vnet_buffer2 (b0)->nat.arc_next;
+      ip0 =
+       (ip4_header_t *) ((u8 *) vlib_buffer_get_current (b0) + iph_offset0);
 
-         ip0 = (ip4_header_t *) ((u8 *) vlib_buffer_get_current (b0) +
-                                 iph_offset0);
+      sw_if_index0 = vnet_buffer (b0)->sw_if_index[VLIB_RX];
+      rx_fib_index0 =
+       fib_table_get_index_for_sw_if_index (FIB_PROTOCOL_IP4, sw_if_index0);
 
-         sw_if_index0 = vnet_buffer (b0)->sw_if_index[VLIB_RX];
-         rx_fib_index0 =
-           fib_table_get_index_for_sw_if_index (FIB_PROTOCOL_IP4,
-                                                sw_if_index0);
+      if (PREDICT_FALSE (ip0->ttl == 1))
+       {
+         vnet_buffer (b0)->sw_if_index[VLIB_TX] = (u32) ~ 0;
+         icmp4_error_set_vnet_buffer (b0, ICMP4_time_exceeded,
+                                      ICMP4_time_exceeded_ttl_exceeded_in_transit,
+                                      0);
+         next[0] = NAT_NEXT_ICMP_ERROR;
+         goto trace0;
+       }
 
-         if (PREDICT_FALSE (ip0->ttl == 1))
-           {
-             vnet_buffer (b0)->sw_if_index[VLIB_TX] = (u32) ~ 0;
-             icmp4_error_set_vnet_buffer (b0, ICMP4_time_exceeded,
-                                          ICMP4_time_exceeded_ttl_exceeded_in_transit,
-                                          0);
-             next0 = NAT_NEXT_ICMP_ERROR;
-             goto trace0;
-           }
+      udp0 = ip4_next_header (ip0);
+      tcp0 = (tcp_header_t *) udp0;
+      proto0 = ip_proto_to_nat_proto (ip0->protocol);
 
-         udp0 = ip4_next_header (ip0);
-         tcp0 = (tcp_header_t *) udp0;
-         proto0 = ip_proto_to_nat_proto (ip0->protocol);
+      if (PREDICT_FALSE (proto0 == NAT_PROTOCOL_OTHER))
+       {
+         next[0] = def_slow;
+         goto trace0;
+       }
 
-         if (PREDICT_FALSE (proto0 == NAT_PROTOCOL_OTHER))
-           {
-             next0 = def_slow;
-             goto trace0;
-           }
+      if (is_output_feature)
+       {
+         if (PREDICT_FALSE
+             (nat_not_translate_output_feature_fwd
+              (sm, ip0, thread_index, now, vm, b0)))
+           goto trace0;
+       }
 
-         if (is_output_feature)
-           {
-             if (PREDICT_FALSE (nat_not_translate_output_feature_fwd
-                                (sm, ip0, thread_index, now, vm, b0)))
-               goto trace0;
-           }
+      if (PREDICT_FALSE (proto0 == NAT_PROTOCOL_ICMP))
+       {
+         next[0] = def_slow;
+         goto trace0;
+       }
 
-         if (PREDICT_FALSE (proto0 == NAT_PROTOCOL_ICMP))
-           {
-             next0 = def_slow;
-             goto trace0;
-           }
+      init_ed_k (&kv0, ip0->src_address,
+                vnet_buffer (b0)->ip.reass.l4_src_port, ip0->dst_address,
+                vnet_buffer (b0)->ip.reass.l4_dst_port, rx_fib_index0,
+                ip0->protocol);
 
-         init_ed_k (&kv0, ip0->src_address,
-                    vnet_buffer (b0)->ip.reass.l4_src_port, ip0->dst_address,
-                    vnet_buffer (b0)->ip.reass.l4_dst_port, rx_fib_index0,
-                    ip0->protocol);
+      // lookup for session
+      if (clib_bihash_search_16_8 (&tsm->in2out_ed, &kv0, &value0))
+       {
+         // session does not exist go slow path
+         next[0] = def_slow;
+         goto trace0;
+       }
+      ASSERT (thread_index == ed_value_get_thread_index (&value0));
+      s0 =
+       pool_elt_at_index (tsm->sessions,
+                          ed_value_get_session_index (&value0));
 
-         // lookup for session
-         if (clib_bihash_search_16_8 (&tsm->in2out_ed, &kv0, &value0))
-           {
-             // session does not exist go slow path
-             next0 = def_slow;
-             goto trace0;
-           }
-         ASSERT (thread_index == ed_value_get_thread_index (&value0));
-         s0 =
-           pool_elt_at_index (tsm->sessions,
-                              ed_value_get_session_index (&value0));
+      if (PREDICT_FALSE (per_vrf_sessions_is_expired (s0, thread_index)))
+       {
+         // session is closed, go slow path
+         nat_free_session_data (sm, s0, thread_index, 0);
+         nat_ed_session_delete (sm, s0, thread_index, 1);
+         next[0] = NAT_NEXT_OUT2IN_ED_SLOW_PATH;
+         goto trace0;
+       }
 
-         if (s0->tcp_closed_timestamp)
+      if (s0->tcp_closed_timestamp)
+       {
+         if (now >= s0->tcp_closed_timestamp)
            {
-             if (now >= s0->tcp_closed_timestamp)
-               {
-                 // session is closed, go slow path
-                 next0 = def_slow;
-               }
-             else
-               {
-                 // session in transitory timeout, drop
-                 b0->error = node->errors[NAT_IN2OUT_ED_ERROR_TCP_CLOSED];
-                 next0 = NAT_NEXT_DROP;
-               }
-             goto trace0;
+             // session is closed, go slow path, freed in slow path
+             next[0] = def_slow;
            }
-
-         // drop if session expired
-         u64 sess_timeout_time;
-         sess_timeout_time = s0->last_heard +
-           (f64) nat44_session_get_timeout (sm, s0);
-         if (now >= sess_timeout_time)
+         else
            {
-             nat_free_session_data (sm, s0, thread_index, 0);
-             nat_ed_session_delete (sm, s0, thread_index, 1);
-             // session is closed, go slow path
-             next0 = def_slow;
-             goto trace0;
+             // session in transitory timeout, drop
+             b0->error = node->errors[NAT_IN2OUT_ED_ERROR_TCP_CLOSED];
+             next[0] = NAT_NEXT_DROP;
            }
+         goto trace0;
+       }
 
-         b0->flags |= VNET_BUFFER_F_IS_NATED;
+      // drop if session expired
+      u64 sess_timeout_time;
+      sess_timeout_time =
+       s0->last_heard + (f64) nat44_session_get_timeout (sm, s0);
+      if (now >= sess_timeout_time)
+       {
+         nat_free_session_data (sm, s0, thread_index, 0);
+         nat_ed_session_delete (sm, s0, thread_index, 1);
+         // session is closed, go slow path
+         next[0] = def_slow;
+         goto trace0;
+       }
 
-         if (!is_output_feature)
-           vnet_buffer (b0)->sw_if_index[VLIB_TX] = s0->out2in.fib_index;
+      b0->flags |= VNET_BUFFER_F_IS_NATED;
 
-         old_addr0 = ip0->src_address.as_u32;
-         new_addr0 = ip0->src_address.as_u32 = s0->out2in.addr.as_u32;
-         sum0 = ip0->checksum;
-         sum0 = ip_csum_update (sum0, old_addr0, new_addr0, ip4_header_t,
-                                src_address);
-         if (PREDICT_FALSE (is_twice_nat_session (s0)))
-           sum0 = ip_csum_update (sum0, ip0->dst_address.as_u32,
-                                  s0->ext_host_addr.as_u32, ip4_header_t,
-                                  dst_address);
-         ip0->checksum = ip_csum_fold (sum0);
+      if (!is_output_feature)
+       vnet_buffer (b0)->sw_if_index[VLIB_TX] = s0->out2in.fib_index;
 
-         old_port0 = vnet_buffer (b0)->ip.reass.l4_src_port;
+      old_addr0 = ip0->src_address.as_u32;
+      new_addr0 = ip0->src_address.as_u32 = s0->out2in.addr.as_u32;
+      sum0 = ip0->checksum;
+      sum0 =
+       ip_csum_update (sum0, old_addr0, new_addr0, ip4_header_t,
+                       src_address);
+      if (PREDICT_FALSE (is_twice_nat_session (s0)))
+       sum0 =
+         ip_csum_update (sum0, ip0->dst_address.as_u32,
+                         s0->ext_host_addr.as_u32, ip4_header_t,
+                         dst_address);
+      ip0->checksum = ip_csum_fold (sum0);
 
-         if (PREDICT_TRUE (proto0 == NAT_PROTOCOL_TCP))
-           {
-             if (!vnet_buffer (b0)->ip.reass.is_non_first_fragment)
-               {
-                 new_port0 = udp0->src_port = s0->out2in.port;
-                 sum0 = tcp0->checksum;
-                 sum0 =
-                   ip_csum_update (sum0, old_addr0, new_addr0,
-                                   ip4_header_t, dst_address);
-                 sum0 =
-                   ip_csum_update (sum0, old_port0, new_port0,
-                                   ip4_header_t, length);
-                 if (PREDICT_FALSE (is_twice_nat_session (s0)))
-                   {
-                     sum0 =
-                       ip_csum_update (sum0, ip0->dst_address.as_u32,
-                                       s0->ext_host_addr.as_u32,
-                                       ip4_header_t, dst_address);
-                     sum0 =
-                       ip_csum_update (sum0,
-                                       vnet_buffer (b0)->ip.
-                                       reass.l4_dst_port, s0->ext_host_port,
-                                       ip4_header_t, length);
-                     tcp0->dst_port = s0->ext_host_port;
-                     ip0->dst_address.as_u32 = s0->ext_host_addr.as_u32;
-                   }
-                 mss_clamping (sm->mss_clamping, tcp0, &sum0);
-                 tcp0->checksum = ip_csum_fold (sum0);
-               }
-             tcp_packets++;
-             if (nat44_set_tcp_session_state_i2o
-                 (sm, now, s0, b0, thread_index))
-               goto trace0;
-           }
-         else if (!vnet_buffer (b0)->ip.reass.is_non_first_fragment
-                  && udp0->checksum)
+      old_port0 = vnet_buffer (b0)->ip.reass.l4_src_port;
+
+      if (PREDICT_TRUE (proto0 == NAT_PROTOCOL_TCP))
+       {
+         if (!vnet_buffer (b0)->ip.reass.is_non_first_fragment)
            {
              new_port0 = udp0->src_port = s0->out2in.port;
-             sum0 = udp0->checksum;
+             sum0 = tcp0->checksum;
              sum0 =
                ip_csum_update (sum0, old_addr0, new_addr0, ip4_header_t,
                                dst_address);
@@ -1117,79 +1100,104 @@ nat44_ed_in2out_fast_path_node_fn_inline (vlib_main_t * vm,
                                length);
              if (PREDICT_FALSE (is_twice_nat_session (s0)))
                {
-                 sum0 = ip_csum_update (sum0, ip0->dst_address.as_u32,
-                                        s0->ext_host_addr.as_u32,
-                                        ip4_header_t, dst_address);
+                 sum0 =
+                   ip_csum_update (sum0, ip0->dst_address.as_u32,
+                                   s0->ext_host_addr.as_u32, ip4_header_t,
+                                   dst_address);
                  sum0 =
                    ip_csum_update (sum0,
                                    vnet_buffer (b0)->ip.reass.l4_dst_port,
                                    s0->ext_host_port, ip4_header_t, length);
-                 udp0->dst_port = s0->ext_host_port;
+                 tcp0->dst_port = s0->ext_host_port;
                  ip0->dst_address.as_u32 = s0->ext_host_addr.as_u32;
                }
-             udp0->checksum = ip_csum_fold (sum0);
-             udp_packets++;
+             mss_clamping (sm->mss_clamping, tcp0, &sum0);
+             tcp0->checksum = ip_csum_fold (sum0);
            }
-         else
+         vlib_increment_simple_counter (&sm->counters.fastpath.in2out_ed.tcp,
+                                        thread_index, sw_if_index0, 1);
+         if (nat44_set_tcp_session_state_i2o (sm, now, s0, b0, thread_index))
+           goto trace0;
+       }
+      else if (!vnet_buffer (b0)->ip.reass.is_non_first_fragment
+              && udp0->checksum)
+       {
+         new_port0 = udp0->src_port = s0->out2in.port;
+         sum0 = udp0->checksum;
+         sum0 =
+           ip_csum_update (sum0, old_addr0, new_addr0, ip4_header_t,
+                           dst_address);
+         sum0 =
+           ip_csum_update (sum0, old_port0, new_port0, ip4_header_t, length);
+         if (PREDICT_FALSE (is_twice_nat_session (s0)))
+           {
+             sum0 =
+               ip_csum_update (sum0, ip0->dst_address.as_u32,
+                               s0->ext_host_addr.as_u32, ip4_header_t,
+                               dst_address);
+             sum0 =
+               ip_csum_update (sum0, vnet_buffer (b0)->ip.reass.l4_dst_port,
+                               s0->ext_host_port, ip4_header_t, length);
+             udp0->dst_port = s0->ext_host_port;
+             ip0->dst_address.as_u32 = s0->ext_host_addr.as_u32;
+           }
+         udp0->checksum = ip_csum_fold (sum0);
+         vlib_increment_simple_counter (&sm->counters.fastpath.in2out_ed.udp,
+                                        thread_index, sw_if_index0, 1);
+       }
+      else
+       {
+         if (!vnet_buffer (b0)->ip.reass.is_non_first_fragment)
            {
-             if (!vnet_buffer (b0)->ip.reass.is_non_first_fragment)
+             new_port0 = udp0->src_port = s0->out2in.port;
+             if (PREDICT_FALSE (is_twice_nat_session (s0)))
                {
-                 new_port0 = udp0->src_port = s0->out2in.port;
-                 if (PREDICT_FALSE (is_twice_nat_session (s0)))
-                   {
-                     udp0->dst_port = s0->ext_host_port;
-                     ip0->dst_address.as_u32 = s0->ext_host_addr.as_u32;
-                   }
-                 udp_packets++;
+                 udp0->dst_port = s0->ext_host_port;
+                 ip0->dst_address.as_u32 = s0->ext_host_addr.as_u32;
                }
+             vlib_increment_simple_counter (&sm->counters.fastpath.
+                                            in2out_ed.udp, thread_index,
+                                            sw_if_index0, 1);
            }
+       }
 
-         /* Accounting */
-         nat44_session_update_counters (s0, now,
-                                        vlib_buffer_length_in_chain
-                                        (vm, b0), thread_index);
-         /* Per-user LRU list maintenance */
-         nat44_session_update_lru (sm, s0, thread_index);
+      /* Accounting */
+      nat44_session_update_counters (s0, now,
+                                    vlib_buffer_length_in_chain (vm, b0),
+                                    thread_index);
+      /* Per-user LRU list maintenance */
+      nat44_session_update_lru (sm, s0, thread_index);
 
-       trace0:
-         if (PREDICT_FALSE ((node->flags & VLIB_NODE_FLAG_TRACE)
-                            && (b0->flags & VLIB_BUFFER_IS_TRACED)))
-           {
-             nat_in2out_ed_trace_t *t =
-               vlib_add_trace (vm, node, b0, sizeof (*t));
-             t->sw_if_index = sw_if_index0;
-             t->next_index = next0;
-             t->is_slow_path = 0;
-
-             if (s0)
-               t->session_index = s0 - tsm->sessions;
-             else
-               t->session_index = ~0;
-           }
+    trace0:
+      if (PREDICT_FALSE
+         ((node->flags & VLIB_NODE_FLAG_TRACE)
+          && (b0->flags & VLIB_BUFFER_IS_TRACED)))
+       {
+         nat_in2out_ed_trace_t *t =
+           vlib_add_trace (vm, node, b0, sizeof (*t));
+         t->sw_if_index = sw_if_index0;
+         t->next_index = next[0];
+         t->is_slow_path = 0;
+
+         if (s0)
+           t->session_index = s0 - tsm->sessions;
+         else
+           t->session_index = ~0;
+       }
 
-         pkts_processed += next0 == vnet_buffer2 (b0)->nat.arc_next;
-         /* verify speculative enqueue, maybe switch current next frame */
-         vlib_validate_buffer_enqueue_x1 (vm, node, next_index,
-                                          to_next, n_left_to_next,
-                                          bi0, next0);
+      if (next[0] == NAT_NEXT_DROP)
+       {
+         vlib_increment_simple_counter (&sm->counters.fastpath.
+                                        in2out_ed.drops, thread_index,
+                                        sw_if_index0, 1);
        }
 
-      vlib_put_next_frame (vm, node, next_index, n_left_to_next);
+      n_left_from--;
+      next++;
     }
 
-  vlib_node_increment_counter (vm, stats_node_index,
-                              NAT_IN2OUT_ED_ERROR_IN2OUT_PACKETS,
-                              pkts_processed);
-  vlib_node_increment_counter (vm, stats_node_index,
-                              NAT_IN2OUT_ED_ERROR_TCP_PACKETS, tcp_packets);
-  vlib_node_increment_counter (vm, stats_node_index,
-                              NAT_IN2OUT_ED_ERROR_UDP_PACKETS, udp_packets);
-  vlib_node_increment_counter (vm, stats_node_index,
-                              NAT_IN2OUT_ED_ERROR_ICMP_PACKETS,
-                              icmp_packets);
-  vlib_node_increment_counter (vm, stats_node_index,
-                              NAT_IN2OUT_ED_ERROR_OTHER_PACKETS,
-                              other_packets);
+  vlib_buffer_enqueue_to_next (vm, node, from, (u16 *) nexts,
+                              frame->n_vectors);
   return frame->n_vectors;
 }
 
@@ -1199,306 +1207,284 @@ nat44_ed_in2out_slow_path_node_fn_inline (vlib_main_t * vm,
                                          vlib_frame_t * frame,
                                          int is_output_feature)
 {
-  u32 n_left_from, *from, *to_next, pkts_processed = 0, stats_node_index;
-  nat_next_t next_index;
+  u32 n_left_from, *from;
   snat_main_t *sm = &snat_main;
   f64 now = vlib_time_now (vm);
   u32 thread_index = vm->thread_index;
   snat_main_per_thread_data_t *tsm = &sm->per_thread_data[thread_index];
-  u32 tcp_packets = 0, udp_packets = 0, icmp_packets = 0, other_packets = 0;
-
-  stats_node_index = sm->ed_in2out_slowpath_node_index;
 
   from = vlib_frame_vector_args (frame);
   n_left_from = frame->n_vectors;
-  next_index = node->cached_next_index;
+
+  vlib_buffer_t *bufs[VLIB_FRAME_SIZE], **b = bufs;
+  u16 nexts[VLIB_FRAME_SIZE], *next = nexts;
+  vlib_get_buffers (vm, from, b, n_left_from);
 
   while (n_left_from > 0)
     {
-      u32 n_left_to_next;
+      vlib_buffer_t *b0;
+      u32 sw_if_index0, rx_fib_index0, proto0, new_addr0, old_addr0,
+       iph_offset0 = 0;
+      u16 old_port0, new_port0;
+      ip4_header_t *ip0;
+      udp_header_t *udp0;
+      tcp_header_t *tcp0;
+      icmp46_header_t *icmp0;
+      snat_session_t *s0 = 0;
+      clib_bihash_kv_16_8_t kv0, value0;
+      ip_csum_t sum0;
 
-      vlib_get_next_frame (vm, node, next_index, to_next, n_left_to_next);
+      b0 = *b;
 
-      while (n_left_from > 0 && n_left_to_next > 0)
-       {
-         u32 bi0;
-         vlib_buffer_t *b0;
-         u32 next0, sw_if_index0, rx_fib_index0, iph_offset0 = 0, proto0,
-           new_addr0, old_addr0;
-         u16 old_port0, new_port0;
-         ip4_header_t *ip0;
-         udp_header_t *udp0;
-         tcp_header_t *tcp0;
-         icmp46_header_t *icmp0;
-         snat_session_t *s0 = 0;
-         clib_bihash_kv_16_8_t kv0, value0;
-         ip_csum_t sum0;
-
-         /* speculatively enqueue b0 to the current next frame */
-         bi0 = from[0];
-         to_next[0] = bi0;
-         from += 1;
-         to_next += 1;
-         n_left_from -= 1;
-         n_left_to_next -= 1;
-
-         b0 = vlib_get_buffer (vm, bi0);
+      if (is_output_feature)
+       iph_offset0 = vnet_buffer (b0)->ip.reass.save_rewrite_length;
 
-         if (is_output_feature)
-           iph_offset0 = vnet_buffer (b0)->ip.reass.save_rewrite_length;
+      next[0] = vnet_buffer2 (b0)->nat.arc_next;
+
+      ip0 = (ip4_header_t *) ((u8 *) vlib_buffer_get_current (b0) +
+                             iph_offset0);
 
-         next0 = vnet_buffer2 (b0)->nat.arc_next;
+      sw_if_index0 = vnet_buffer (b0)->sw_if_index[VLIB_RX];
+      rx_fib_index0 =
+       fib_table_get_index_for_sw_if_index (FIB_PROTOCOL_IP4, sw_if_index0);
 
-         ip0 = (ip4_header_t *) ((u8 *) vlib_buffer_get_current (b0) +
-                                 iph_offset0);
+      if (PREDICT_FALSE (ip0->ttl == 1))
+       {
+         vnet_buffer (b0)->sw_if_index[VLIB_TX] = (u32) ~ 0;
+         icmp4_error_set_vnet_buffer (b0, ICMP4_time_exceeded,
+                                      ICMP4_time_exceeded_ttl_exceeded_in_transit,
+                                      0);
+         next[0] = NAT_NEXT_ICMP_ERROR;
+         goto trace0;
+       }
 
-         sw_if_index0 = vnet_buffer (b0)->sw_if_index[VLIB_RX];
-         rx_fib_index0 =
-           fib_table_get_index_for_sw_if_index (FIB_PROTOCOL_IP4,
-                                                sw_if_index0);
+      udp0 = ip4_next_header (ip0);
+      tcp0 = (tcp_header_t *) udp0;
+      icmp0 = (icmp46_header_t *) udp0;
+      proto0 = ip_proto_to_nat_proto (ip0->protocol);
 
-         if (PREDICT_FALSE (ip0->ttl == 1))
-           {
-             vnet_buffer (b0)->sw_if_index[VLIB_TX] = (u32) ~ 0;
-             icmp4_error_set_vnet_buffer (b0, ICMP4_time_exceeded,
-                                          ICMP4_time_exceeded_ttl_exceeded_in_transit,
-                                          0);
-             next0 = NAT_NEXT_ICMP_ERROR;
-             goto trace0;
-           }
+      if (PREDICT_FALSE (proto0 == NAT_PROTOCOL_OTHER))
+       {
+         s0 = nat44_ed_in2out_unknown_proto (sm, b0, ip0,
+                                             rx_fib_index0,
+                                             thread_index, now, vm, node);
+         if (!s0)
+           next[0] = NAT_NEXT_DROP;
 
-         udp0 = ip4_next_header (ip0);
-         tcp0 = (tcp_header_t *) udp0;
-         icmp0 = (icmp46_header_t *) udp0;
-         proto0 = ip_proto_to_nat_proto (ip0->protocol);
+         vlib_increment_simple_counter (&sm->counters.slowpath.
+                                        in2out_ed.other, thread_index,
+                                        sw_if_index0, 1);
+         goto trace0;
+       }
 
-         if (PREDICT_FALSE (proto0 == NAT_PROTOCOL_OTHER))
-           {
-             s0 = nat44_ed_in2out_unknown_proto (sm, b0, ip0,
-                                                 rx_fib_index0,
-                                                 thread_index, now,
-                                                 vm, node);
-             if (!s0)
-               next0 = NAT_NEXT_DROP;
-
-             other_packets++;
-             goto trace0;
-           }
+      if (PREDICT_FALSE (proto0 == NAT_PROTOCOL_ICMP))
+       {
+         next[0] =
+           icmp_in2out_ed_slow_path (sm, b0, ip0, icmp0, sw_if_index0,
+                                     rx_fib_index0, node, next[0], now,
+                                     thread_index, &s0);
+         vlib_increment_simple_counter (&sm->counters.slowpath.
+                                        in2out_ed.icmp, thread_index,
+                                        sw_if_index0, 1);
+         goto trace0;
+       }
+
+      init_ed_k (&kv0, ip0->src_address,
+                vnet_buffer (b0)->ip.reass.l4_src_port, ip0->dst_address,
+                vnet_buffer (b0)->ip.reass.l4_dst_port, rx_fib_index0,
+                ip0->protocol);
+      if (!clib_bihash_search_16_8 (&tsm->in2out_ed, &kv0, &value0))
+       {
+         ASSERT (thread_index == ed_value_get_thread_index (&value0));
+         s0 =
+           pool_elt_at_index (tsm->sessions,
+                              ed_value_get_session_index (&value0));
 
-         if (PREDICT_FALSE (proto0 == NAT_PROTOCOL_ICMP))
+         if (s0->tcp_closed_timestamp && now >= s0->tcp_closed_timestamp)
            {
-             next0 = icmp_in2out_ed_slow_path
-               (sm, b0, ip0, icmp0, sw_if_index0, rx_fib_index0,
-                node, next0, now, thread_index, &s0);
-             icmp_packets++;
-             goto trace0;
+             nat_free_session_data (sm, s0, thread_index, 0);
+             nat_ed_session_delete (sm, s0, thread_index, 1);
+             s0 = NULL;
            }
+       }
 
-         init_ed_k (&kv0, ip0->src_address,
-                    vnet_buffer (b0)->ip.reass.l4_src_port, ip0->dst_address,
-                    vnet_buffer (b0)->ip.reass.l4_dst_port, rx_fib_index0,
-                    ip0->protocol);
-         if (!clib_bihash_search_16_8 (&tsm->in2out_ed, &kv0, &value0))
+      if (!s0)
+       {
+         if (is_output_feature)
            {
-             ASSERT (thread_index == ed_value_get_thread_index (&value0));
-             s0 =
-               pool_elt_at_index (tsm->sessions,
-                                  ed_value_get_session_index (&value0));
+             if (PREDICT_FALSE
+                 (nat44_ed_not_translate_output_feature
+                  (sm, ip0, vnet_buffer (b0)->ip.reass.l4_src_port,
+                   vnet_buffer (b0)->ip.reass.l4_dst_port, thread_index,
+                   sw_if_index0, vnet_buffer (b0)->sw_if_index[VLIB_TX])))
+               goto trace0;
 
-             if (s0->tcp_closed_timestamp && now >= s0->tcp_closed_timestamp)
-               {
-                 nat_free_session_data (sm, s0, thread_index, 0);
-                 nat_ed_session_delete (sm, s0, thread_index, 1);
-                 s0 = NULL;
-               }
+             /*
+              * Send DHCP packets to the ipv4 stack, or we won't
+              * be able to use dhcp client on the outside interface
+              */
+             if (PREDICT_FALSE
+                 (proto0 == NAT_PROTOCOL_UDP
+                  && (vnet_buffer (b0)->ip.reass.l4_dst_port ==
+                      clib_host_to_net_u16 (UDP_DST_PORT_dhcp_to_server))
+                  && ip0->dst_address.as_u32 == 0xffffffff))
+               goto trace0;
            }
-
-         if (!s0)
+         else
            {
-             if (is_output_feature)
-               {
-                 if (PREDICT_FALSE
-                     (nat44_ed_not_translate_output_feature
-                      (sm, ip0, vnet_buffer (b0)->ip.reass.l4_src_port,
-                       vnet_buffer (b0)->ip.reass.l4_dst_port, thread_index,
-                       sw_if_index0,
-                       vnet_buffer (b0)->sw_if_index[VLIB_TX])))
-                   goto trace0;
-
-                 /*
-                  * Send DHCP packets to the ipv4 stack, or we won't
-                  * be able to use dhcp client on the outside interface
-                  */
-                 if (PREDICT_FALSE
-                     (proto0 == NAT_PROTOCOL_UDP
-                      && (vnet_buffer (b0)->ip.reass.l4_dst_port ==
-                          clib_host_to_net_u16 (UDP_DST_PORT_dhcp_to_server))
-                      && ip0->dst_address.as_u32 == 0xffffffff))
-                   goto trace0;
-               }
-             else
-               {
-                 if (PREDICT_FALSE
-                     (nat44_ed_not_translate
-                      (sm, node, sw_if_index0, ip0, proto0, rx_fib_index0,
-                       thread_index)))
-                   goto trace0;
-               }
+             if (PREDICT_FALSE
+                 (nat44_ed_not_translate
+                  (sm, node, sw_if_index0, ip0, proto0, rx_fib_index0,
+                   thread_index)))
+               goto trace0;
+           }
 
-             next0 =
-               slow_path_ed (sm, b0, ip0->src_address, ip0->dst_address,
-                             vnet_buffer (b0)->ip.reass.l4_src_port,
-                             vnet_buffer (b0)->ip.reass.l4_dst_port,
-                             ip0->protocol, rx_fib_index0, &s0, node, next0,
-                             thread_index, now);
+         next[0] =
+           slow_path_ed (sm, b0, ip0->src_address, ip0->dst_address,
+                         vnet_buffer (b0)->ip.reass.l4_src_port,
+                         vnet_buffer (b0)->ip.reass.l4_dst_port,
+                         ip0->protocol, rx_fib_index0, &s0, node, next[0],
+                         thread_index, now);
 
-             if (PREDICT_FALSE (next0 == NAT_NEXT_DROP))
-               goto trace0;
+         if (PREDICT_FALSE (next[0] == NAT_NEXT_DROP))
+           goto trace0;
 
-             if (PREDICT_FALSE (!s0))
-               goto trace0;
+         if (PREDICT_FALSE (!s0))
+           goto trace0;
 
-           }
+       }
 
-         b0->flags |= VNET_BUFFER_F_IS_NATED;
+      b0->flags |= VNET_BUFFER_F_IS_NATED;
 
-         if (!is_output_feature)
-           vnet_buffer (b0)->sw_if_index[VLIB_TX] = s0->out2in.fib_index;
+      if (!is_output_feature)
+       vnet_buffer (b0)->sw_if_index[VLIB_TX] = s0->out2in.fib_index;
 
-         old_addr0 = ip0->src_address.as_u32;
-         new_addr0 = ip0->src_address.as_u32 = s0->out2in.addr.as_u32;
-         sum0 = ip0->checksum;
-         sum0 = ip_csum_update (sum0, old_addr0, new_addr0, ip4_header_t,
-                                src_address);
-         if (PREDICT_FALSE (is_twice_nat_session (s0)))
-           sum0 = ip_csum_update (sum0, ip0->dst_address.as_u32,
-                                  s0->ext_host_addr.as_u32, ip4_header_t,
-                                  dst_address);
-         ip0->checksum = ip_csum_fold (sum0);
+      old_addr0 = ip0->src_address.as_u32;
+      new_addr0 = ip0->src_address.as_u32 = s0->out2in.addr.as_u32;
+      sum0 = ip0->checksum;
+      sum0 = ip_csum_update (sum0, old_addr0, new_addr0, ip4_header_t,
+                            src_address);
+      if (PREDICT_FALSE (is_twice_nat_session (s0)))
+       sum0 = ip_csum_update (sum0, ip0->dst_address.as_u32,
+                              s0->ext_host_addr.as_u32, ip4_header_t,
+                              dst_address);
+      ip0->checksum = ip_csum_fold (sum0);
 
-         old_port0 = vnet_buffer (b0)->ip.reass.l4_src_port;
+      old_port0 = vnet_buffer (b0)->ip.reass.l4_src_port;
 
-         if (PREDICT_TRUE (proto0 == NAT_PROTOCOL_TCP))
-           {
-             if (!vnet_buffer (b0)->ip.reass.is_non_first_fragment)
-               {
-                 new_port0 = udp0->src_port = s0->out2in.port;
-                 sum0 = tcp0->checksum;
-                 sum0 =
-                   ip_csum_update (sum0, old_addr0, new_addr0,
-                                   ip4_header_t, dst_address);
-                 sum0 =
-                   ip_csum_update (sum0, old_port0, new_port0,
-                                   ip4_header_t, length);
-                 if (PREDICT_FALSE (is_twice_nat_session (s0)))
-                   {
-                     sum0 =
-                       ip_csum_update (sum0, ip0->dst_address.as_u32,
-                                       s0->ext_host_addr.as_u32,
-                                       ip4_header_t, dst_address);
-                     sum0 =
-                       ip_csum_update (sum0,
-                                       vnet_buffer (b0)->ip.
-                                       reass.l4_dst_port, s0->ext_host_port,
-                                       ip4_header_t, length);
-                     tcp0->dst_port = s0->ext_host_port;
-                     ip0->dst_address.as_u32 = s0->ext_host_addr.as_u32;
-                   }
-                 mss_clamping (sm->mss_clamping, tcp0, &sum0);
-                 tcp0->checksum = ip_csum_fold (sum0);
-               }
-             tcp_packets++;
-             if (nat44_set_tcp_session_state_i2o
-                 (sm, now, s0, b0, thread_index))
-               goto trace0;
-           }
-         else if (!vnet_buffer (b0)->ip.reass.is_non_first_fragment
-                  && udp0->checksum)
+      if (PREDICT_TRUE (proto0 == NAT_PROTOCOL_TCP))
+       {
+         if (!vnet_buffer (b0)->ip.reass.is_non_first_fragment)
            {
              new_port0 = udp0->src_port = s0->out2in.port;
-             sum0 = udp0->checksum;
+             sum0 = tcp0->checksum;
              sum0 =
-               ip_csum_update (sum0, old_addr0, new_addr0, ip4_header_t,
-                               dst_address);
+               ip_csum_update (sum0, old_addr0, new_addr0,
+                               ip4_header_t, dst_address);
              sum0 =
-               ip_csum_update (sum0, old_port0, new_port0, ip4_header_t,
-                               length);
+               ip_csum_update (sum0, old_port0, new_port0,
+                               ip4_header_t, length);
              if (PREDICT_FALSE (is_twice_nat_session (s0)))
                {
-                 sum0 = ip_csum_update (sum0, ip0->dst_address.as_u32,
-                                        s0->ext_host_addr.as_u32,
-                                        ip4_header_t, dst_address);
+                 sum0 =
+                   ip_csum_update (sum0, ip0->dst_address.as_u32,
+                                   s0->ext_host_addr.as_u32,
+                                   ip4_header_t, dst_address);
                  sum0 =
                    ip_csum_update (sum0,
                                    vnet_buffer (b0)->ip.reass.l4_dst_port,
                                    s0->ext_host_port, ip4_header_t, length);
-                 udp0->dst_port = s0->ext_host_port;
+                 tcp0->dst_port = s0->ext_host_port;
                  ip0->dst_address.as_u32 = s0->ext_host_addr.as_u32;
                }
-             udp0->checksum = ip_csum_fold (sum0);
-             udp_packets++;
+             mss_clamping (sm->mss_clamping, tcp0, &sum0);
+             tcp0->checksum = ip_csum_fold (sum0);
            }
-         else
+         vlib_increment_simple_counter (&sm->counters.slowpath.in2out_ed.tcp,
+                                        thread_index, sw_if_index0, 1);
+         if (nat44_set_tcp_session_state_i2o (sm, now, s0, b0, thread_index))
+           goto trace0;
+       }
+      else if (!vnet_buffer (b0)->ip.reass.is_non_first_fragment
+              && udp0->checksum)
+       {
+         new_port0 = udp0->src_port = s0->out2in.port;
+         sum0 = udp0->checksum;
+         sum0 =
+           ip_csum_update (sum0, old_addr0, new_addr0, ip4_header_t,
+                           dst_address);
+         sum0 =
+           ip_csum_update (sum0, old_port0, new_port0, ip4_header_t, length);
+         if (PREDICT_FALSE (is_twice_nat_session (s0)))
+           {
+             sum0 = ip_csum_update (sum0, ip0->dst_address.as_u32,
+                                    s0->ext_host_addr.as_u32,
+                                    ip4_header_t, dst_address);
+             sum0 =
+               ip_csum_update (sum0,
+                               vnet_buffer (b0)->ip.reass.l4_dst_port,
+                               s0->ext_host_port, ip4_header_t, length);
+             udp0->dst_port = s0->ext_host_port;
+             ip0->dst_address.as_u32 = s0->ext_host_addr.as_u32;
+           }
+         udp0->checksum = ip_csum_fold (sum0);
+         vlib_increment_simple_counter (&sm->counters.slowpath.in2out_ed.udp,
+                                        thread_index, sw_if_index0, 1);
+       }
+      else
+       {
+         if (!vnet_buffer (b0)->ip.reass.is_non_first_fragment)
            {
-             if (!vnet_buffer (b0)->ip.reass.is_non_first_fragment)
+             new_port0 = udp0->src_port = s0->out2in.port;
+             if (PREDICT_FALSE (is_twice_nat_session (s0)))
                {
-                 new_port0 = udp0->src_port = s0->out2in.port;
-                 if (PREDICT_FALSE (is_twice_nat_session (s0)))
-                   {
-                     udp0->dst_port = s0->ext_host_port;
-                     ip0->dst_address.as_u32 = s0->ext_host_addr.as_u32;
-                   }
-                 udp_packets++;
+                 udp0->dst_port = s0->ext_host_port;
+                 ip0->dst_address.as_u32 = s0->ext_host_addr.as_u32;
                }
+             vlib_increment_simple_counter (&sm->counters.slowpath.
+                                            in2out_ed.udp, thread_index,
+                                            sw_if_index0, 1);
            }
+       }
 
-         /* Accounting */
-         nat44_session_update_counters (s0, now,
-                                        vlib_buffer_length_in_chain
-                                        (vm, b0), thread_index);
-         /* Per-user LRU list maintenance */
-         nat44_session_update_lru (sm, s0, thread_index);
-
-       trace0:
-         if (PREDICT_FALSE ((node->flags & VLIB_NODE_FLAG_TRACE)
-                            && (b0->flags & VLIB_BUFFER_IS_TRACED)))
-           {
-             nat_in2out_ed_trace_t *t =
-               vlib_add_trace (vm, node, b0, sizeof (*t));
-             t->sw_if_index = sw_if_index0;
-             t->next_index = next0;
-             t->is_slow_path = 1;
-
-             if (s0)
-               t->session_index = s0 - tsm->sessions;
-             else
-               t->session_index = ~0;
-           }
+      /* Accounting */
+      nat44_session_update_counters (s0, now,
+                                    vlib_buffer_length_in_chain
+                                    (vm, b0), thread_index);
+      /* Per-user LRU list maintenance */
+      nat44_session_update_lru (sm, s0, thread_index);
 
-         pkts_processed += next0 == vnet_buffer2 (b0)->nat.arc_next;
+    trace0:
+      if (PREDICT_FALSE ((node->flags & VLIB_NODE_FLAG_TRACE)
+                        && (b0->flags & VLIB_BUFFER_IS_TRACED)))
+       {
+         nat_in2out_ed_trace_t *t =
+           vlib_add_trace (vm, node, b0, sizeof (*t));
+         t->sw_if_index = sw_if_index0;
+         t->next_index = next[0];
+         t->is_slow_path = 1;
+
+         if (s0)
+           t->session_index = s0 - tsm->sessions;
+         else
+           t->session_index = ~0;
+       }
 
-         /* verify speculative enqueue, maybe switch current next frame */
-         vlib_validate_buffer_enqueue_x1 (vm, node, next_index,
-                                          to_next, n_left_to_next,
-                                          bi0, next0);
+      if (next[0] == NAT_NEXT_DROP)
+       {
+         vlib_increment_simple_counter (&sm->counters.slowpath.
+                                        in2out_ed.drops, thread_index,
+                                        sw_if_index0, 1);
        }
 
-      vlib_put_next_frame (vm, node, next_index, n_left_to_next);
+      n_left_from--;
+      next++;
+      b++;
     }
 
-  vlib_node_increment_counter (vm, stats_node_index,
-                              NAT_IN2OUT_ED_ERROR_IN2OUT_PACKETS,
-                              pkts_processed);
-  vlib_node_increment_counter (vm, stats_node_index,
-                              NAT_IN2OUT_ED_ERROR_TCP_PACKETS, tcp_packets);
-  vlib_node_increment_counter (vm, stats_node_index,
-                              NAT_IN2OUT_ED_ERROR_UDP_PACKETS, udp_packets);
-  vlib_node_increment_counter (vm, stats_node_index,
-                              NAT_IN2OUT_ED_ERROR_ICMP_PACKETS,
-                              icmp_packets);
-  vlib_node_increment_counter (vm, stats_node_index,
-                              NAT_IN2OUT_ED_ERROR_OTHER_PACKETS,
-                              other_packets);
+  vlib_buffer_enqueue_to_next (vm, node, from, (u16 *) nexts,
+                              frame->n_vectors);
+
   return frame->n_vectors;
 }