nat: use SVR
[vpp.git] / src / plugins / nat / nat44_handoff.c
index d7ab650..d221e5b 100644 (file)
 #include <vnet/fib/ip4_fib.h>
 #include <vppinfra/error.h>
 #include <nat/nat.h>
+#include <nat/nat_inlines.h>
 
 typedef struct
 {
   u32 next_worker_index;
+  u32 trace_index;
   u8 in2out;
+  u8 output;
 } nat44_handoff_trace_t;
 
 #define foreach_nat44_handoff_error                       \
@@ -50,101 +53,208 @@ static char *nat44_handoff_error_strings[] = {
 };
 
 
-vlib_node_registration_t snat_in2out_worker_handoff_node;
-vlib_node_registration_t snat_in2out_output_worker_handoff_node;
-vlib_node_registration_t snat_out2in_worker_handoff_node;
-
 static u8 *
 format_nat44_handoff_trace (u8 * s, va_list * args)
 {
   CLIB_UNUSED (vlib_main_t * vm) = va_arg (*args, vlib_main_t *);
   CLIB_UNUSED (vlib_node_t * node) = va_arg (*args, vlib_node_t *);
   nat44_handoff_trace_t *t = va_arg (*args, nat44_handoff_trace_t *);
-  char *tag;
+  char *tag, *output;
 
   tag = t->in2out ? "IN2OUT" : "OUT2IN";
+  output = t->output ? "OUTPUT-FEATURE" : "";
   s =
-    format (s, "NAT44_%s_WORKER_HANDOFF: next-worker %d", tag,
-           t->next_worker_index);
+    format (s, "NAT44_%s_WORKER_HANDOFF %s: next-worker %d trace index %d",
+           tag, output, t->next_worker_index, t->trace_index);
 
   return s;
 }
 
 static inline uword
-nat44_worker_handoff_fn_inline (vlib_main_t * vm, vlib_node_runtime_t * node,
+nat44_worker_handoff_fn_inline (vlib_main_t * vm,
+                               vlib_node_runtime_t * node,
                                vlib_frame_t * frame, u8 is_output,
                                u8 is_in2out)
 {
+  u32 n_enq, n_left_from, *from, do_handoff = 0, same_worker = 0;
+
+  u16 thread_indices[VLIB_FRAME_SIZE], *ti = thread_indices;
+  vlib_buffer_t *bufs[VLIB_FRAME_SIZE], **b = bufs;
   snat_main_t *sm = &snat_main;
-  vlib_buffer_t *bufs[VLIB_FRAME_SIZE], **b;
-  u32 n_enq, n_left_from, *from;
-  u16 thread_indices[VLIB_FRAME_SIZE], *ti;
-  u32 fq_index;
-  snat_get_worker_function_t *get_worker;
-  u32 thread_index = vm->thread_index;
-  u32 do_handoff = 0, same_worker = 0;
+
+  u32 fq_index, thread_index = vm->thread_index;
 
   from = vlib_frame_vector_args (frame);
   n_left_from = frame->n_vectors;
-  vlib_get_buffers (vm, from, bufs, n_left_from);
 
-  b = bufs;
-  ti = thread_indices;
-
-  ASSERT (vec_len (sm->workers));
+  vlib_get_buffers (vm, from, b, n_left_from);
 
   if (is_in2out)
     {
-      get_worker = sm->worker_in2out_cb;
-      if (is_output)
-       fq_index = sm->fq_in2out_output_index;
-      else
-       fq_index = sm->fq_in2out_index;
+      fq_index = is_output ? sm->fq_in2out_output_index : sm->fq_in2out_index;
     }
   else
     {
       fq_index = sm->fq_out2in_index;
-      get_worker = sm->worker_out2in_cb;
+    }
+
+  while (n_left_from >= 4)
+    {
+      u32 sw_if_index0, sw_if_index1, sw_if_index2, sw_if_index3;
+      u32 rx_fib_index0, rx_fib_index1, rx_fib_index2, rx_fib_index3;
+      u32 iph_offset0 = 0, iph_offset1 = 0, iph_offset2 = 0, iph_offset3 = 0;
+      ip4_header_t *ip0, *ip1, *ip2, *ip3;
+
+      if (PREDICT_TRUE (n_left_from >= 8))
+       {
+         vlib_prefetch_buffer_header (b[4], STORE);
+         vlib_prefetch_buffer_header (b[5], STORE);
+         vlib_prefetch_buffer_header (b[6], STORE);
+         vlib_prefetch_buffer_header (b[7], STORE);
+         CLIB_PREFETCH (&b[4]->data, CLIB_CACHE_LINE_BYTES, STORE);
+         CLIB_PREFETCH (&b[5]->data, CLIB_CACHE_LINE_BYTES, STORE);
+         CLIB_PREFETCH (&b[6]->data, CLIB_CACHE_LINE_BYTES, STORE);
+         CLIB_PREFETCH (&b[7]->data, CLIB_CACHE_LINE_BYTES, STORE);
+       }
+
+      if (is_output)
+       {
+         iph_offset0 = vnet_buffer (b[0])->ip.save_rewrite_length;
+         iph_offset1 = vnet_buffer (b[1])->ip.save_rewrite_length;
+         iph_offset2 = vnet_buffer (b[2])->ip.save_rewrite_length;
+         iph_offset3 = vnet_buffer (b[3])->ip.save_rewrite_length;
+       }
+
+      ip0 = (ip4_header_t *) ((u8 *) vlib_buffer_get_current (b[0]) +
+                             iph_offset0);
+      ip1 = (ip4_header_t *) ((u8 *) vlib_buffer_get_current (b[1]) +
+                             iph_offset1);
+      ip2 = (ip4_header_t *) ((u8 *) vlib_buffer_get_current (b[2]) +
+                             iph_offset2);
+      ip3 = (ip4_header_t *) ((u8 *) vlib_buffer_get_current (b[3]) +
+                             iph_offset3);
+
+      sw_if_index0 = vnet_buffer (b[0])->sw_if_index[VLIB_RX];
+      sw_if_index1 = vnet_buffer (b[1])->sw_if_index[VLIB_RX];
+      sw_if_index2 = vnet_buffer (b[2])->sw_if_index[VLIB_RX];
+      sw_if_index3 = vnet_buffer (b[3])->sw_if_index[VLIB_RX];
+
+      rx_fib_index0 = ip4_fib_table_get_index_for_sw_if_index (sw_if_index0);
+      rx_fib_index1 = ip4_fib_table_get_index_for_sw_if_index (sw_if_index1);
+      rx_fib_index2 = ip4_fib_table_get_index_for_sw_if_index (sw_if_index2);
+      rx_fib_index3 = ip4_fib_table_get_index_for_sw_if_index (sw_if_index3);
+
+      if (is_in2out)
+       {
+         ti[0] = sm->worker_in2out_cb (ip0, rx_fib_index0, is_output);
+         ti[1] = sm->worker_in2out_cb (ip1, rx_fib_index1, is_output);
+         ti[2] = sm->worker_in2out_cb (ip2, rx_fib_index2, is_output);
+         ti[3] = sm->worker_in2out_cb (ip3, rx_fib_index3, is_output);
+       }
+      else
+       {
+         ti[0] = sm->worker_out2in_cb (b[0], ip0, rx_fib_index0, is_output);
+         ti[1] = sm->worker_out2in_cb (b[1], ip1, rx_fib_index1, is_output);
+         ti[2] = sm->worker_out2in_cb (b[2], ip2, rx_fib_index2, is_output);
+         ti[3] = sm->worker_out2in_cb (b[3], ip3, rx_fib_index3, is_output);
+       }
+
+      if (ti[0] == thread_index)
+       same_worker++;
+      else
+       do_handoff++;
+
+      if (ti[1] == thread_index)
+       same_worker++;
+      else
+       do_handoff++;
+
+      if (ti[2] == thread_index)
+       same_worker++;
+      else
+       do_handoff++;
+
+      if (ti[3] == thread_index)
+       same_worker++;
+      else
+       do_handoff++;
+
+      b += 4;
+      ti += 4;
+      n_left_from -= 4;
     }
 
   while (n_left_from > 0)
     {
       u32 sw_if_index0;
       u32 rx_fib_index0;
+      u32 iph_offset0 = 0;
       ip4_header_t *ip0;
 
+
+      if (is_output)
+       iph_offset0 = vnet_buffer (b[0])->ip.save_rewrite_length;
+
+      ip0 = (ip4_header_t *) ((u8 *) vlib_buffer_get_current (b[0]) +
+                             iph_offset0);
+
       sw_if_index0 = vnet_buffer (b[0])->sw_if_index[VLIB_RX];
       rx_fib_index0 = ip4_fib_table_get_index_for_sw_if_index (sw_if_index0);
-      ip0 = vlib_buffer_get_current (b[0]);
-      ti[0] = get_worker (ip0, rx_fib_index0);
 
-      if (ti[0] != thread_index)
-       do_handoff++;
+      if (is_in2out)
+       {
+         ti[0] = sm->worker_in2out_cb (ip0, rx_fib_index0, is_output);
+       }
       else
-       same_worker++;
-
-      if (PREDICT_FALSE ((node->flags & VLIB_NODE_FLAG_TRACE)
-                        && (b[0]->flags & VLIB_BUFFER_IS_TRACED)))
        {
-         nat44_handoff_trace_t *t =
-           vlib_add_trace (vm, node, b[0], sizeof (*t));
-         t->next_worker_index = ti[0];
-         t->in2out = is_in2out;
+         ti[0] = sm->worker_out2in_cb (b[0], ip0, rx_fib_index0, is_output);
        }
 
-      n_left_from -= 1;
-      ti += 1;
+      if (ti[0] == thread_index)
+       same_worker++;
+      else
+       do_handoff++;
+
       b += 1;
+      ti += 1;
+      n_left_from -= 1;
+    }
+
+  if (PREDICT_FALSE ((node->flags & VLIB_NODE_FLAG_TRACE)))
+    {
+      u32 i;
+      b = bufs;
+      ti = thread_indices;
+
+      for (i = 0; i < frame->n_vectors; i++)
+       {
+         if (b[0]->flags & VLIB_BUFFER_IS_TRACED)
+           {
+             nat44_handoff_trace_t *t =
+               vlib_add_trace (vm, node, b[0], sizeof (*t));
+             t->next_worker_index = ti[0];
+             t->trace_index = vlib_buffer_get_trace_index (b[0]);
+             t->in2out = is_in2out;
+             t->output = is_output;
+
+             b += 1;
+             ti += 1;
+           }
+         else
+           break;
+       }
     }
 
-  n_enq =
-    vlib_buffer_enqueue_to_thread (vm, fq_index, from, thread_indices,
-                                  frame->n_vectors, 1);
+  n_enq = vlib_buffer_enqueue_to_thread (vm, fq_index, from, thread_indices,
+                                        frame->n_vectors, 1);
 
   if (n_enq < frame->n_vectors)
-    vlib_node_increment_counter (vm, node->node_index,
-                                NAT44_HANDOFF_ERROR_CONGESTION_DROP,
-                                frame->n_vectors - n_enq);
+    {
+      vlib_node_increment_counter (vm, node->node_index,
+                                  NAT44_HANDOFF_ERROR_CONGESTION_DROP,
+                                  frame->n_vectors - n_enq);
+    }
+
   vlib_node_increment_counter (vm, node->node_index,
                               NAT44_HANDOFF_ERROR_SAME_WORKER, same_worker);
   vlib_node_increment_counter (vm, node->node_index,
@@ -152,87 +262,66 @@ nat44_worker_handoff_fn_inline (vlib_main_t * vm, vlib_node_runtime_t * node,
   return frame->n_vectors;
 }
 
-static uword
-snat_in2out_worker_handoff_fn (vlib_main_t * vm,
-                              vlib_node_runtime_t * node,
-                              vlib_frame_t * frame)
+
+
+VLIB_NODE_FN (snat_in2out_worker_handoff_node) (vlib_main_t * vm,
+                                               vlib_node_runtime_t * node,
+                                               vlib_frame_t * frame)
 {
   return nat44_worker_handoff_fn_inline (vm, node, frame, 0, 1);
 }
 
 /* *INDENT-OFF* */
 VLIB_REGISTER_NODE (snat_in2out_worker_handoff_node) = {
-  .function = snat_in2out_worker_handoff_fn,
   .name = "nat44-in2out-worker-handoff",
   .vector_size = sizeof (u32),
+  .sibling_of = "nat-default",
   .format_trace = format_nat44_handoff_trace,
   .type = VLIB_NODE_TYPE_INTERNAL,
   .n_errors = ARRAY_LEN(nat44_handoff_error_strings),
   .error_strings = nat44_handoff_error_strings,
-  .n_next_nodes = 1,
-  .next_nodes = {
-    [0] = "error-drop",
-  },
 };
 /* *INDENT-ON* */
 
-VLIB_NODE_FUNCTION_MULTIARCH (snat_in2out_worker_handoff_node,
-                             snat_in2out_worker_handoff_fn);
-
-static uword
-snat_in2out_output_worker_handoff_fn (vlib_main_t * vm,
-                                     vlib_node_runtime_t * node,
-                                     vlib_frame_t * frame)
+VLIB_NODE_FN (snat_in2out_output_worker_handoff_node) (vlib_main_t * vm,
+                                                      vlib_node_runtime_t *
+                                                      node,
+                                                      vlib_frame_t * frame)
 {
   return nat44_worker_handoff_fn_inline (vm, node, frame, 1, 1);
 }
 
 /* *INDENT-OFF* */
 VLIB_REGISTER_NODE (snat_in2out_output_worker_handoff_node) = {
-  .function = snat_in2out_output_worker_handoff_fn,
   .name = "nat44-in2out-output-worker-handoff",
   .vector_size = sizeof (u32),
+  .sibling_of = "nat-default",
   .format_trace = format_nat44_handoff_trace,
   .type = VLIB_NODE_TYPE_INTERNAL,
   .n_errors = ARRAY_LEN(nat44_handoff_error_strings),
   .error_strings = nat44_handoff_error_strings,
-  .n_next_nodes = 1,
-  .next_nodes = {
-    [0] = "error-drop",
-  },
 };
 /* *INDENT-ON* */
 
-VLIB_NODE_FUNCTION_MULTIARCH (snat_in2out_output_worker_handoff_node,
-                             snat_in2out_output_worker_handoff_fn);
-
-static uword
-snat_out2in_worker_handoff_fn (vlib_main_t * vm,
-                              vlib_node_runtime_t * node,
-                              vlib_frame_t * frame)
+VLIB_NODE_FN (snat_out2in_worker_handoff_node) (vlib_main_t * vm,
+                                               vlib_node_runtime_t * node,
+                                               vlib_frame_t * frame)
 {
   return nat44_worker_handoff_fn_inline (vm, node, frame, 0, 0);
 }
 
 /* *INDENT-OFF* */
 VLIB_REGISTER_NODE (snat_out2in_worker_handoff_node) = {
-  .function = snat_out2in_worker_handoff_fn,
   .name = "nat44-out2in-worker-handoff",
   .vector_size = sizeof (u32),
+  .sibling_of = "nat-default",
   .format_trace = format_nat44_handoff_trace,
   .type = VLIB_NODE_TYPE_INTERNAL,
   .n_errors = ARRAY_LEN(nat44_handoff_error_strings),
   .error_strings = nat44_handoff_error_strings,
-  .n_next_nodes = 1,
-  .next_nodes = {
-    [0] = "error-drop",
-  },
 };
 /* *INDENT-ON* */
 
-VLIB_NODE_FUNCTION_MULTIARCH (snat_out2in_worker_handoff_node,
-                             snat_out2in_worker_handoff_fn);
-
 /*
  * fd.io coding-style-patch-verification: ON
  *