gbp: migrate old MULTIARCH macros to VLIB_NODE_FN 65/17965/4
authorFilip Tehlar <ftehlar@cisco.com>
Thu, 28 Feb 2019 10:51:47 +0000 (02:51 -0800)
committerNeale Ranns <nranns@cisco.com>
Mon, 4 Mar 2019 07:49:25 +0000 (07:49 +0000)
Change-Id: I3b10caa447b796172f787df8fcbb92f2b4dd2803
Signed-off-by: Filip Tehlar <ftehlar@cisco.com>
13 files changed:
src/plugins/gbp/CMakeLists.txt
src/plugins/gbp/gbp.h
src/plugins/gbp/gbp_classify.c
src/plugins/gbp/gbp_fwd.c
src/plugins/gbp/gbp_fwd_dpo.c
src/plugins/gbp/gbp_fwd_node.c [new file with mode: 0644]
src/plugins/gbp/gbp_learn.c
src/plugins/gbp/gbp_policy.c
src/plugins/gbp/gbp_policy_dpo.c
src/plugins/gbp/gbp_sclass.c
src/plugins/gbp/gbp_vxlan.c
src/plugins/gbp/gbp_vxlan.h
src/plugins/gbp/gbp_vxlan_node.c [new file with mode: 0644]

index ef25402..140a456 100644 (file)
@@ -22,6 +22,7 @@ add_vpp_plugin(gbp
   gbp_ext_itf.c
   gbp_fwd.c
   gbp_fwd_dpo.c
+  gbp_fwd_node.c
   gbp_itf.c
   gbp_learn.c
   gbp_policy.c
@@ -32,6 +33,17 @@ add_vpp_plugin(gbp
   gbp_sclass.c
   gbp_subnet.c
   gbp_vxlan.c
+  gbp_vxlan_node.c
+
+  MULTIARCH_SOURCES
+  gbp_fwd_node.c
+  gbp_policy.c
+  gbp_learn.c
+  gbp_policy_dpo.c
+  gbp_fwd_dpo.c
+  gbp_vxlan_node.c
+  gbp_sclass.c
+  gbp_classify.c
 
   API_FILES
   gbp.api
index 06612cd..35e02d2 100644 (file)
@@ -48,6 +48,19 @@ typedef struct
 
 extern gbp_main_t gbp_main;
 
+/**
+ * Grouping of global data for the GBP source EPG classification feature
+ */
+typedef struct gbp_policy_main_t_
+{
+  /**
+   * Next nodes for L2 output features
+   */
+  u32 l2_output_feat_next[2][32];
+} gbp_policy_main_t;
+
+extern gbp_policy_main_t gbp_policy_main;
+
 #endif
 
 /*
index ba6fe9b..2eaa7bb 100644 (file)
@@ -185,33 +185,33 @@ gbp_classify_inline (vlib_main_t * vm,
   return frame->n_vectors;
 }
 
-static uword
-gbp_src_classify (vlib_main_t * vm,
-                 vlib_node_runtime_t * node, vlib_frame_t * frame)
+VLIB_NODE_FN (gbp_src_classify_node) (vlib_main_t * vm,
+                                     vlib_node_runtime_t * node,
+                                     vlib_frame_t * frame)
 {
   return (gbp_classify_inline (vm, node, frame,
                               GBP_SRC_CLASSIFY_PORT, DPO_PROTO_ETHERNET));
 }
 
-static uword
-gbp_null_classify (vlib_main_t * vm,
-                  vlib_node_runtime_t * node, vlib_frame_t * frame)
+VLIB_NODE_FN (gbp_null_classify_node) (vlib_main_t * vm,
+                                      vlib_node_runtime_t * node,
+                                      vlib_frame_t * frame)
 {
   return (gbp_classify_inline (vm, node, frame,
                               GBP_SRC_CLASSIFY_NULL, DPO_PROTO_ETHERNET));
 }
 
-static uword
-gbp_ip4_src_classify (vlib_main_t * vm,
-                     vlib_node_runtime_t * node, vlib_frame_t * frame)
+VLIB_NODE_FN (gbp_ip4_src_classify_node) (vlib_main_t * vm,
+                                         vlib_node_runtime_t * node,
+                                         vlib_frame_t * frame)
 {
   return (gbp_classify_inline (vm, node, frame,
                               GBP_SRC_CLASSIFY_PORT, DPO_PROTO_IP4));
 }
 
-static uword
-gbp_ip6_src_classify (vlib_main_t * vm,
-                     vlib_node_runtime_t * node, vlib_frame_t * frame)
+VLIB_NODE_FN (gbp_ip6_src_classify_node) (vlib_main_t * vm,
+                                         vlib_node_runtime_t * node,
+                                         vlib_frame_t * frame)
 {
   return (gbp_classify_inline (vm, node, frame,
                               GBP_SRC_CLASSIFY_PORT, DPO_PROTO_IP6));
@@ -233,7 +233,6 @@ format_gbp_classify_trace (u8 * s, va_list * args)
 
 /* *INDENT-OFF* */
 VLIB_REGISTER_NODE (gbp_null_classify_node) = {
-  .function = gbp_null_classify,
   .name = "gbp-null-classify",
   .vector_size = sizeof (u32),
   .format_trace = format_gbp_classify_trace,
@@ -243,10 +242,7 @@ VLIB_REGISTER_NODE (gbp_null_classify_node) = {
   .n_next_nodes = 0,
 };
 
-VLIB_NODE_FUNCTION_MULTIARCH (gbp_null_classify_node, gbp_null_classify);
-
 VLIB_REGISTER_NODE (gbp_src_classify_node) = {
-  .function = gbp_src_classify,
   .name = "gbp-src-classify",
   .vector_size = sizeof (u32),
   .format_trace = format_gbp_classify_trace,
@@ -256,10 +252,7 @@ VLIB_REGISTER_NODE (gbp_src_classify_node) = {
   .n_next_nodes = 0,
 };
 
-VLIB_NODE_FUNCTION_MULTIARCH (gbp_src_classify_node, gbp_src_classify);
-
 VLIB_REGISTER_NODE (gbp_ip4_src_classify_node) = {
-  .function = gbp_ip4_src_classify,
   .name = "ip4-gbp-src-classify",
   .vector_size = sizeof (u32),
   .format_trace = format_gbp_classify_trace,
@@ -272,10 +265,7 @@ VLIB_REGISTER_NODE (gbp_ip4_src_classify_node) = {
   },
 };
 
-VLIB_NODE_FUNCTION_MULTIARCH (gbp_ip4_src_classify_node, gbp_ip4_src_classify);
-
 VLIB_REGISTER_NODE (gbp_ip6_src_classify_node) = {
-  .function = gbp_ip6_src_classify,
   .name = "ip6-gbp-src-classify",
   .vector_size = sizeof (u32),
   .format_trace = format_gbp_classify_trace,
@@ -288,8 +278,6 @@ VLIB_REGISTER_NODE (gbp_ip6_src_classify_node) = {
   },
 };
 
-VLIB_NODE_FUNCTION_MULTIARCH (gbp_ip6_src_classify_node, gbp_ip6_src_classify);
-
 VNET_FEATURE_INIT (gbp_ip4_src_classify_feat_node, static) =
 {
   .arc_name = "ip4-unicast",
@@ -490,30 +478,29 @@ gbp_lpm_classify_inline (vlib_main_t * vm,
   return frame->n_vectors;
 }
 
-static uword
-gbp_ip4_lpm_classify (vlib_main_t * vm,
-                     vlib_node_runtime_t * node, vlib_frame_t * frame)
+VLIB_NODE_FN (gbp_ip4_lpm_classify_node) (vlib_main_t * vm,
+                                         vlib_node_runtime_t * node,
+                                         vlib_frame_t * frame)
 {
   return (gbp_lpm_classify_inline (vm, node, frame, DPO_PROTO_IP4, 1));
 }
 
-static uword
-gbp_ip6_lpm_classify (vlib_main_t * vm,
-                     vlib_node_runtime_t * node, vlib_frame_t * frame)
+VLIB_NODE_FN (gbp_ip6_lpm_classify_node) (vlib_main_t * vm,
+                                         vlib_node_runtime_t * node,
+                                         vlib_frame_t * frame)
 {
   return (gbp_lpm_classify_inline (vm, node, frame, DPO_PROTO_IP6, 1));
 }
 
-static uword
-gbp_l2_lpm_classify (vlib_main_t * vm,
-                    vlib_node_runtime_t * node, vlib_frame_t * frame)
+VLIB_NODE_FN (gbp_l2_lpm_classify_node) (vlib_main_t * vm,
+                                        vlib_node_runtime_t * node,
+                                        vlib_frame_t * frame)
 {
   return (gbp_lpm_classify_inline (vm, node, frame, DPO_PROTO_ETHERNET, 0));
 }
 
 /* *INDENT-OFF* */
 VLIB_REGISTER_NODE (gbp_ip4_lpm_classify_node) = {
-  .function = gbp_ip4_lpm_classify,
   .name = "ip4-gbp-lpm-classify",
   .vector_size = sizeof (u32),
   .format_trace = format_gbp_classify_trace,
@@ -526,10 +513,7 @@ VLIB_REGISTER_NODE (gbp_ip4_lpm_classify_node) = {
   },
 };
 
-VLIB_NODE_FUNCTION_MULTIARCH (gbp_ip4_lpm_classify_node, gbp_ip4_lpm_classify);
-
 VLIB_REGISTER_NODE (gbp_ip6_lpm_classify_node) = {
-  .function = gbp_ip6_lpm_classify,
   .name = "ip6-gbp-lpm-classify",
   .vector_size = sizeof (u32),
   .format_trace = format_gbp_classify_trace,
@@ -542,10 +526,7 @@ VLIB_REGISTER_NODE (gbp_ip6_lpm_classify_node) = {
   },
 };
 
-VLIB_NODE_FUNCTION_MULTIARCH (gbp_ip6_lpm_classify_node, gbp_ip6_lpm_classify);
-
 VLIB_REGISTER_NODE (gbp_l2_lpm_classify_node) = {
-  .function = gbp_l2_lpm_classify,
   .name = "l2-gbp-lpm-classify",
   .vector_size = sizeof (u32),
   .format_trace = format_gbp_classify_trace,
@@ -558,8 +539,6 @@ VLIB_REGISTER_NODE (gbp_l2_lpm_classify_node) = {
   },
 };
 
-VLIB_NODE_FUNCTION_MULTIARCH (gbp_l2_lpm_classify_node, gbp_l2_lpm_classify);
-
 VNET_FEATURE_INIT (gbp_ip4_lpm_classify_feat_node, static) =
 {
   .arc_name = "ip4-unicast",
@@ -580,19 +559,25 @@ gbp_src_classify_init (vlib_main_t * vm)
 {
   gbp_src_classify_main_t *em = &gbp_src_classify_main;
 
+  vlib_node_t *node = vlib_get_node_by_name (vm, (u8 *) "gbp-src-classify");
+
   /* Initialize the feature next-node indexes */
   feat_bitmap_init_next_nodes (vm,
-                              gbp_src_classify_node.index,
+                              node->index,
                               L2INPUT_N_FEAT,
                               l2input_get_feat_names (),
                               em->l2_input_feat_next[GBP_SRC_CLASSIFY_NULL]);
+
+  node = vlib_get_node_by_name (vm, (u8 *) "gbp-null-classify");
   feat_bitmap_init_next_nodes (vm,
-                              gbp_null_classify_node.index,
+                              node->index,
                               L2INPUT_N_FEAT,
                               l2input_get_feat_names (),
                               em->l2_input_feat_next[GBP_SRC_CLASSIFY_PORT]);
+
+  node = vlib_get_node_by_name (vm, (u8 *) "l2-gbp-lpm-classify");
   feat_bitmap_init_next_nodes (vm,
-                              gbp_l2_lpm_classify_node.index,
+                              node->index,
                               L2INPUT_N_FEAT,
                               l2input_get_feat_names (),
                               em->l2_input_feat_next[GBP_SRC_CLASSIFY_LPM]);
index e481433..8d98f1c 100644 (file)
@@ -1,5 +1,5 @@
 /*
- * Copyright (c) 2018 Cisco and/or its affiliates.
+ * Copyright (c) 2019 Cisco and/or its affiliates.
  * Licensed under the Apache License, Version 2.0 (the "License");
  * you may not use this file except in compliance with the License.
  * You may obtain a copy of the License at:
@@ -12,9 +12,9 @@
  * See the License for the specific language governing permissions and
  * limitations under the License.
  */
-
 #include <plugins/gbp/gbp.h>
 #include <vnet/l2/l2_input.h>
+#include <plugins/gbp/gbp_learn.h>
 
 /**
  * Grouping of global data for the GBP source EPG classification feature
@@ -29,155 +29,49 @@ typedef struct gbp_fwd_main_t_
 
 static gbp_fwd_main_t gbp_fwd_main;
 
-#define foreach_gbp_fwd                      \
-  _(DROP,    "drop")                         \
-  _(OUTPUT,  "output")
-
-typedef enum
-{
-#define _(sym,str) GBP_FWD_ERROR_##sym,
-  foreach_gbp_fwd
-#undef _
-    GBP_FWD_N_ERROR,
-} gbp_fwd_error_t;
-
-static char *gbp_fwd_error_strings[] = {
-#define _(sym,string) string,
-  foreach_gbp_fwd
-#undef _
-};
-
-typedef enum
-{
-#define _(sym,str) GBP_FWD_NEXT_##sym,
-  foreach_gbp_fwd
-#undef _
-    GBP_FWD_N_NEXT,
-} gbp_fwd_next_t;
-
-/**
- * per-packet trace data
- */
-typedef struct gbp_fwd_trace_t_
-{
-  /* per-pkt trace data */
-  epg_id_t src_epg;
-  u32 sw_if_index;
-} gbp_fwd_trace_t;
+gbp_policy_main_t gbp_policy_main;
 
-static uword
-gbp_fwd (vlib_main_t * vm, vlib_node_runtime_t * node, vlib_frame_t * frame)
+void
+gbp_learn_enable (u32 sw_if_index, gbb_learn_mode_t mode)
 {
-  u32 n_left_from, *from, *to_next;
-  u32 next_index;
-
-  next_index = 0;
-  n_left_from = frame->n_vectors;
-  from = vlib_frame_vector_args (frame);
-
-  while (n_left_from > 0)
+  if (GBP_LEARN_MODE_L2 == mode)
     {
-      u32 n_left_to_next;
-
-      vlib_get_next_frame (vm, node, next_index, to_next, n_left_to_next);
-
-      while (n_left_from > 0 && n_left_to_next > 0)
-       {
-         u32 bi0, sw_if_index0, src_epg;
-         gbp_fwd_next_t next0;
-         vlib_buffer_t *b0;
-
-         next0 = GBP_FWD_NEXT_DROP;
-         bi0 = from[0];
-         to_next[0] = bi0;
-         from += 1;
-         to_next += 1;
-         n_left_from -= 1;
-         n_left_to_next -= 1;
-
-         b0 = vlib_get_buffer (vm, bi0);
-
-         /*
-          * lookup the uplink based on src EPG
-          */
-         src_epg = vnet_buffer2 (b0)->gbp.src_epg;
-
-         sw_if_index0 = gbp_epg_itf_lookup (src_epg);
-
-         if (~0 != sw_if_index0)
-           {
-             vnet_buffer (b0)->sw_if_index[VLIB_TX] = sw_if_index0;
-
-             next0 = GBP_FWD_NEXT_OUTPUT;
-           }
-         /*
-          * else
-          *  don't know the uplink interface for this EPG => drop
-          */
-
-         if (PREDICT_FALSE ((b0->flags & VLIB_BUFFER_IS_TRACED)))
-           {
-             gbp_fwd_trace_t *t = vlib_add_trace (vm, node, b0, sizeof (*t));
-             t->src_epg = src_epg;
-             t->sw_if_index = sw_if_index0;
-           }
-
-         /* verify speculative enqueue, maybe switch current next frame */
-         vlib_validate_buffer_enqueue_x1 (vm, node, next_index,
-                                          to_next, n_left_to_next,
-                                          bi0, next0);
-       }
-
-      vlib_put_next_frame (vm, node, next_index, n_left_to_next);
+      l2input_intf_bitmap_enable (sw_if_index, L2INPUT_FEAT_GBP_LEARN, 1);
+    }
+  else
+    {
+      vnet_feature_enable_disable ("ip4-unicast",
+                                  "gbp-learn-ip4", sw_if_index, 1, 0, 0);
+      vnet_feature_enable_disable ("ip6-unicast",
+                                  "gbp-learn-ip6", sw_if_index, 1, 0, 0);
     }
-
-  return frame->n_vectors;
 }
 
-/* packet trace format function */
-static u8 *
-format_gbp_fwd_trace (u8 * s, va_list * args)
+void
+gbp_learn_disable (u32 sw_if_index, gbb_learn_mode_t mode)
 {
-  CLIB_UNUSED (vlib_main_t * vm) = va_arg (*args, vlib_main_t *);
-  CLIB_UNUSED (vlib_node_t * node) = va_arg (*args, vlib_node_t *);
-  gbp_fwd_trace_t *t = va_arg (*args, gbp_fwd_trace_t *);
-
-  s = format (s, "src-epg:%d", t->src_epg);
-
-  return s;
+  if (GBP_LEARN_MODE_L2 == mode)
+    {
+      l2input_intf_bitmap_enable (sw_if_index, L2INPUT_FEAT_GBP_LEARN, 0);
+    }
+  else
+    {
+      vnet_feature_enable_disable ("ip4-unicast",
+                                  "gbp-learn-ip4", sw_if_index, 0, 0, 0);
+      vnet_feature_enable_disable ("ip6-unicast",
+                                  "gbp-learn-ip6", sw_if_index, 0, 0, 0);
+    }
 }
 
-/* *INDENT-OFF* */
-VLIB_REGISTER_NODE (gbp_fwd_node) = {
-  .function = gbp_fwd,
-  .name = "gbp-fwd",
-  .vector_size = sizeof (u32),
-  .format_trace = format_gbp_fwd_trace,
-  .type = VLIB_NODE_TYPE_INTERNAL,
-
-  .n_errors = ARRAY_LEN(gbp_fwd_error_strings),
-  .error_strings = gbp_fwd_error_strings,
-
-  .n_next_nodes = GBP_FWD_N_NEXT,
-
-  .next_nodes = {
-    [GBP_FWD_NEXT_DROP] = "error-drop",
-    [GBP_FWD_NEXT_OUTPUT] = "l2-output",
-  },
-};
-
-VLIB_NODE_FUNCTION_MULTIARCH (gbp_fwd_node, gbp_fwd);
-
-/* *INDENT-ON* */
-
 static clib_error_t *
 gbp_fwd_init (vlib_main_t * vm)
 {
   gbp_fwd_main_t *gpm = &gbp_fwd_main;
+  vlib_node_t *node = vlib_get_node_by_name (vm, (u8 *) "gbp-fwd");
 
   /* Initialize the feature next-node indices */
   feat_bitmap_init_next_nodes (vm,
-                              gbp_fwd_node.index,
+                              node->index,
                               L2INPUT_N_FEAT,
                               l2input_get_feat_names (),
                               gpm->l2_input_feat_next);
index e3dba5f..c0b19a9 100644 (file)
@@ -18,6 +18,8 @@
 
 #include <vnet/ethernet/ethernet.h>
 
+
+#ifndef CLIB_MARCH_VARIANT
 /**
  * The 'DB' of GBP FWD DPOs.
  * There is one per-proto
@@ -164,6 +166,7 @@ gbp_fwd_dpo_module_init (vlib_main_t * vm)
 }
 
 VLIB_INIT_FUNCTION (gbp_fwd_dpo_module_init);
+#endif /* CLIB_MARCH_VARIANT */
 
 typedef struct gbp_fwd_dpo_trace_t_
 {
@@ -255,23 +258,22 @@ format_gbp_fwd_dpo_trace (u8 * s, va_list * args)
   return s;
 }
 
-static uword
-ip4_gbp_fwd_dpo (vlib_main_t * vm,
-                vlib_node_runtime_t * node, vlib_frame_t * from_frame)
+VLIB_NODE_FN (ip4_gbp_fwd_dpo_node) (vlib_main_t * vm,
+                                    vlib_node_runtime_t * node,
+                                    vlib_frame_t * from_frame)
 {
   return (gbp_fwd_dpo_inline (vm, node, from_frame, FIB_PROTOCOL_IP4));
 }
 
-static uword
-ip6_gbp_fwd_dpo (vlib_main_t * vm,
-                vlib_node_runtime_t * node, vlib_frame_t * from_frame)
+VLIB_NODE_FN (ip6_gbp_fwd_dpo_node) (vlib_main_t * vm,
+                                    vlib_node_runtime_t * node,
+                                    vlib_frame_t * from_frame)
 {
   return (gbp_fwd_dpo_inline (vm, node, from_frame, FIB_PROTOCOL_IP6));
 }
 
 /* *INDENT-OFF* */
 VLIB_REGISTER_NODE (ip4_gbp_fwd_dpo_node) = {
-    .function = ip4_gbp_fwd_dpo,
     .name = "ip4-gbp-fwd-dpo",
     .vector_size = sizeof (u32),
     .format_trace = format_gbp_fwd_dpo_trace,
@@ -283,7 +285,6 @@ VLIB_REGISTER_NODE (ip4_gbp_fwd_dpo_node) = {
     }
 };
 VLIB_REGISTER_NODE (ip6_gbp_fwd_dpo_node) = {
-    .function = ip6_gbp_fwd_dpo,
     .name = "ip6-gbp-fwd-dpo",
     .vector_size = sizeof (u32),
     .format_trace = format_gbp_fwd_dpo_trace,
@@ -294,9 +295,6 @@ VLIB_REGISTER_NODE (ip6_gbp_fwd_dpo_node) = {
         [GBP_FWD_FWD] = "ip6-dvr-dpo",
     }
 };
-
-VLIB_NODE_FUNCTION_MULTIARCH (ip4_gbp_fwd_dpo_node, ip4_gbp_fwd_dpo)
-VLIB_NODE_FUNCTION_MULTIARCH (ip6_gbp_fwd_dpo_node, ip6_gbp_fwd_dpo)
 /* *INDENT-ON* */
 
 /*
diff --git a/src/plugins/gbp/gbp_fwd_node.c b/src/plugins/gbp/gbp_fwd_node.c
new file mode 100644 (file)
index 0000000..eff4aeb
--- /dev/null
@@ -0,0 +1,162 @@
+/*
+ * Copyright (c) 2018 Cisco and/or its affiliates.
+ * Licensed under the Apache License, Version 2.0 (the "License");
+ * you may not use this file except in compliance with the License.
+ * You may obtain a copy of the License at:
+ *
+ *     http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+#include <plugins/gbp/gbp.h>
+#include <vnet/l2/l2_input.h>
+
+#define foreach_gbp_fwd                      \
+  _(DROP,    "drop")                         \
+  _(OUTPUT,  "output")
+
+typedef enum
+{
+#define _(sym,str) GBP_FWD_ERROR_##sym,
+  foreach_gbp_fwd
+#undef _
+    GBP_FWD_N_ERROR,
+} gbp_fwd_error_t;
+
+static char *gbp_fwd_error_strings[] = {
+#define _(sym,string) string,
+  foreach_gbp_fwd
+#undef _
+};
+
+typedef enum
+{
+#define _(sym,str) GBP_FWD_NEXT_##sym,
+  foreach_gbp_fwd
+#undef _
+    GBP_FWD_N_NEXT,
+} gbp_fwd_next_t;
+
+/**
+ * per-packet trace data
+ */
+typedef struct gbp_fwd_trace_t_
+{
+  /* per-pkt trace data */
+  epg_id_t src_epg;
+  u32 sw_if_index;
+} gbp_fwd_trace_t;
+
+VLIB_NODE_FN (gbp_fwd_node) (vlib_main_t * vm, vlib_node_runtime_t * node,
+                            vlib_frame_t * frame)
+{
+  u32 n_left_from, *from, *to_next;
+  u32 next_index;
+
+  next_index = 0;
+  n_left_from = frame->n_vectors;
+  from = vlib_frame_vector_args (frame);
+
+  while (n_left_from > 0)
+    {
+      u32 n_left_to_next;
+
+      vlib_get_next_frame (vm, node, next_index, to_next, n_left_to_next);
+
+      while (n_left_from > 0 && n_left_to_next > 0)
+       {
+         u32 bi0, sw_if_index0, src_epg;
+         gbp_fwd_next_t next0;
+         vlib_buffer_t *b0;
+
+         next0 = GBP_FWD_NEXT_DROP;
+         bi0 = from[0];
+         to_next[0] = bi0;
+         from += 1;
+         to_next += 1;
+         n_left_from -= 1;
+         n_left_to_next -= 1;
+
+         b0 = vlib_get_buffer (vm, bi0);
+
+         /*
+          * lookup the uplink based on src EPG
+          */
+         src_epg = vnet_buffer2 (b0)->gbp.src_epg;
+
+         sw_if_index0 = gbp_epg_itf_lookup (src_epg);
+
+         if (~0 != sw_if_index0)
+           {
+             vnet_buffer (b0)->sw_if_index[VLIB_TX] = sw_if_index0;
+
+             next0 = GBP_FWD_NEXT_OUTPUT;
+           }
+         /*
+          * else
+          *  don't know the uplink interface for this EPG => drop
+          */
+
+         if (PREDICT_FALSE ((b0->flags & VLIB_BUFFER_IS_TRACED)))
+           {
+             gbp_fwd_trace_t *t = vlib_add_trace (vm, node, b0, sizeof (*t));
+             t->src_epg = src_epg;
+             t->sw_if_index = sw_if_index0;
+           }
+
+         /* verify speculative enqueue, maybe switch current next frame */
+         vlib_validate_buffer_enqueue_x1 (vm, node, next_index,
+                                          to_next, n_left_to_next,
+                                          bi0, next0);
+       }
+
+      vlib_put_next_frame (vm, node, next_index, n_left_to_next);
+    }
+
+  return frame->n_vectors;
+}
+
+/* packet trace format function */
+static u8 *
+format_gbp_fwd_trace (u8 * s, va_list * args)
+{
+  CLIB_UNUSED (vlib_main_t * vm) = va_arg (*args, vlib_main_t *);
+  CLIB_UNUSED (vlib_node_t * node) = va_arg (*args, vlib_node_t *);
+  gbp_fwd_trace_t *t = va_arg (*args, gbp_fwd_trace_t *);
+
+  s = format (s, "src-epg:%d", t->src_epg);
+
+  return s;
+}
+
+/* *INDENT-OFF* */
+VLIB_REGISTER_NODE (gbp_fwd_node) = {
+  .name = "gbp-fwd",
+  .vector_size = sizeof (u32),
+  .format_trace = format_gbp_fwd_trace,
+  .type = VLIB_NODE_TYPE_INTERNAL,
+
+  .n_errors = ARRAY_LEN(gbp_fwd_error_strings),
+  .error_strings = gbp_fwd_error_strings,
+
+  .n_next_nodes = GBP_FWD_N_NEXT,
+
+  .next_nodes = {
+    [GBP_FWD_NEXT_DROP] = "error-drop",
+    [GBP_FWD_NEXT_OUTPUT] = "l2-output",
+  },
+};
+/* *INDENT-ON* */
+
+/*
+ * fd.io coding-style-patch-verification: ON
+ *
+ * Local Variables:
+ * eval: (c-set-style "gnu")
+ * End:
+ */
index 514aca2..156ea35 100644 (file)
@@ -211,9 +211,9 @@ gbp_learn_get_outer (const ethernet_header_t * eh0,
   *outer_dst = ip0->dst_address;
 }
 
-static uword
-gbp_learn_l2 (vlib_main_t * vm,
-             vlib_node_runtime_t * node, vlib_frame_t * frame)
+VLIB_NODE_FN (gbp_learn_l2_node) (vlib_main_t * vm,
+                                 vlib_node_runtime_t * node,
+                                 vlib_frame_t * frame)
 {
   u32 n_left_from, *from, *to_next, next_index, thread_index, seed;
   gbp_learn_main_t *glm;
@@ -379,7 +379,6 @@ format_gbp_learn_l2_trace (u8 * s, va_list * args)
 
 /* *INDENT-OFF* */
 VLIB_REGISTER_NODE (gbp_learn_l2_node) = {
-  .function = gbp_learn_l2,
   .name = "gbp-learn-l2",
   .vector_size = sizeof (u32),
   .format_trace = format_gbp_learn_l2_trace,
@@ -394,8 +393,6 @@ VLIB_REGISTER_NODE (gbp_learn_l2_node) = {
     [GBP_LEARN_NEXT_DROP] = "error-drop",
   },
 };
-
-VLIB_NODE_FUNCTION_MULTIARCH (gbp_learn_l2_node, gbp_learn_l2);
 /* *INDENT-ON* */
 
 typedef struct gbp_learn_l3_t_
@@ -653,31 +650,28 @@ format_gbp_learn_l3_trace (u8 * s, va_list * args)
   return s;
 }
 
-static uword
-gbp_learn_ip4 (vlib_main_t * vm,
-              vlib_node_runtime_t * node, vlib_frame_t * frame)
+VLIB_NODE_FN (gbp_learn_ip4_node) (vlib_main_t * vm,
+                                  vlib_node_runtime_t * node,
+                                  vlib_frame_t * frame)
 {
   return (gbp_learn_l3 (vm, node, frame, FIB_PROTOCOL_IP4));
 }
 
-static uword
-gbp_learn_ip6 (vlib_main_t * vm,
-              vlib_node_runtime_t * node, vlib_frame_t * frame)
+VLIB_NODE_FN (gbp_learn_ip6_node) (vlib_main_t * vm,
+                                  vlib_node_runtime_t * node,
+                                  vlib_frame_t * frame)
 {
   return (gbp_learn_l3 (vm, node, frame, FIB_PROTOCOL_IP6));
 }
 
 /* *INDENT-OFF* */
 VLIB_REGISTER_NODE (gbp_learn_ip4_node) = {
-  .function = gbp_learn_ip4,
   .name = "gbp-learn-ip4",
   .vector_size = sizeof (u32),
   .format_trace = format_gbp_learn_l3_trace,
   .type = VLIB_NODE_TYPE_INTERNAL,
 };
 
-VLIB_NODE_FUNCTION_MULTIARCH (gbp_learn_ip4_node, gbp_learn_ip4);
-
 VNET_FEATURE_INIT (gbp_learn_ip4, static) =
 {
   .arc_name = "ip4-unicast",
@@ -685,15 +679,12 @@ VNET_FEATURE_INIT (gbp_learn_ip4, static) =
 };
 
 VLIB_REGISTER_NODE (gbp_learn_ip6_node) = {
-  .function = gbp_learn_ip6,
   .name = "gbp-learn-ip6",
   .vector_size = sizeof (u32),
   .format_trace = format_gbp_learn_l3_trace,
   .type = VLIB_NODE_TYPE_INTERNAL,
 };
 
-VLIB_NODE_FUNCTION_MULTIARCH (gbp_learn_ip6_node, gbp_learn_ip6);
-
 VNET_FEATURE_INIT (gbp_learn_ip6, static) =
 {
   .arc_name = "ip6-unicast",
@@ -702,47 +693,17 @@ VNET_FEATURE_INIT (gbp_learn_ip6, static) =
 
 /* *INDENT-ON* */
 
-void
-gbp_learn_enable (u32 sw_if_index, gbb_learn_mode_t mode)
-{
-  if (GBP_LEARN_MODE_L2 == mode)
-    {
-      l2input_intf_bitmap_enable (sw_if_index, L2INPUT_FEAT_GBP_LEARN, 1);
-    }
-  else
-    {
-      vnet_feature_enable_disable ("ip4-unicast",
-                                  "gbp-learn-ip4", sw_if_index, 1, 0, 0);
-      vnet_feature_enable_disable ("ip6-unicast",
-                                  "gbp-learn-ip6", sw_if_index, 1, 0, 0);
-    }
-}
-
-void
-gbp_learn_disable (u32 sw_if_index, gbb_learn_mode_t mode)
-{
-  if (GBP_LEARN_MODE_L2 == mode)
-    {
-      l2input_intf_bitmap_enable (sw_if_index, L2INPUT_FEAT_GBP_LEARN, 0);
-    }
-  else
-    {
-      vnet_feature_enable_disable ("ip4-unicast",
-                                  "gbp-learn-ip4", sw_if_index, 0, 0, 0);
-      vnet_feature_enable_disable ("ip6-unicast",
-                                  "gbp-learn-ip6", sw_if_index, 0, 0, 0);
-    }
-}
-
 static clib_error_t *
 gbp_learn_init (vlib_main_t * vm)
 {
   gbp_learn_main_t *glm = &gbp_learn_main;
   vlib_thread_main_t *tm = &vlib_thread_main;
 
+  vlib_node_t *node = vlib_get_node_by_name (vm, (u8 *) "gbp-learn-l2");
+
   /* Initialize the feature next-node indices */
   feat_bitmap_init_next_nodes (vm,
-                              gbp_learn_l2_node.index,
+                              node->index,
                               L2INPUT_N_FEAT,
                               l2input_get_feat_names (),
                               glm->gl_l2_input_feat_next);
index 7b7bd5c..5d5dcd0 100644 (file)
 
 #include <vnet/vxlan-gbp/vxlan_gbp_packet.h>
 
-/**
- * Grouping of global data for the GBP source EPG classification feature
- */
-typedef struct gbp_policy_main_t_
-{
-  /**
-   * Next nodes for L2 output features
-   */
-  u32 l2_output_feat_next[2][32];
-} gbp_policy_main_t;
-
-static gbp_policy_main_t gbp_policy_main;
-
 #define foreach_gbp_policy                      \
   _(DENY,    "deny")
 
@@ -336,16 +323,16 @@ gbp_policy_inline (vlib_main_t * vm,
   return frame->n_vectors;
 }
 
-static uword
-gbp_policy_port (vlib_main_t * vm,
-                vlib_node_runtime_t * node, vlib_frame_t * frame)
+VLIB_NODE_FN (gbp_policy_port_node) (vlib_main_t * vm,
+                                    vlib_node_runtime_t * node,
+                                    vlib_frame_t * frame)
 {
   return (gbp_policy_inline (vm, node, frame, 1));
 }
 
-static uword
-gbp_policy_mac (vlib_main_t * vm,
-               vlib_node_runtime_t * node, vlib_frame_t * frame)
+VLIB_NODE_FN (gbp_policy_mac_node) (vlib_main_t * vm,
+                                   vlib_node_runtime_t * node,
+                                   vlib_frame_t * frame)
 {
   return (gbp_policy_inline (vm, node, frame, 0));
 }
@@ -367,7 +354,6 @@ format_gbp_policy_trace (u8 * s, va_list * args)
 
 /* *INDENT-OFF* */
 VLIB_REGISTER_NODE (gbp_policy_port_node) = {
-  .function = gbp_policy_port,
   .name = "gbp-policy-port",
   .vector_size = sizeof (u32),
   .format_trace = format_gbp_policy_trace,
@@ -383,10 +369,7 @@ VLIB_REGISTER_NODE (gbp_policy_port_node) = {
   },
 };
 
-VLIB_NODE_FUNCTION_MULTIARCH (gbp_policy_port_node, gbp_policy_port);
-
 VLIB_REGISTER_NODE (gbp_policy_mac_node) = {
-  .function = gbp_policy_mac,
   .name = "gbp-policy-mac",
   .vector_size = sizeof (u32),
   .format_trace = format_gbp_policy_trace,
@@ -394,8 +377,6 @@ VLIB_REGISTER_NODE (gbp_policy_mac_node) = {
   .sibling_of = "gbp-policy-port",
 };
 
-VLIB_NODE_FUNCTION_MULTIARCH (gbp_policy_mac_node, gbp_policy_mac);
-
 /* *INDENT-ON* */
 
 static clib_error_t *
@@ -404,14 +385,18 @@ gbp_policy_init (vlib_main_t * vm)
   gbp_policy_main_t *gpm = &gbp_policy_main;
   clib_error_t *error = 0;
 
+  vlib_node_t *node = vlib_get_node_by_name (vm, (u8 *) "gbp-policy-port");
+
   /* Initialize the feature next-node indexes */
   feat_bitmap_init_next_nodes (vm,
-                              gbp_policy_port_node.index,
+                              node->index,
                               L2OUTPUT_N_FEAT,
                               l2output_get_feat_names (),
                               gpm->l2_output_feat_next[1]);
+
+  node = vlib_get_node_by_name (vm, (u8 *) "gbp-policy-mac");
   feat_bitmap_init_next_nodes (vm,
-                              gbp_policy_mac_node.index,
+                              node->index,
                               L2OUTPUT_N_FEAT,
                               l2output_get_feat_names (),
                               gpm->l2_output_feat_next[0]);
index 213475f..ce628e5 100644 (file)
@@ -21,6 +21,7 @@
 #include <plugins/gbp/gbp_policy_dpo.h>
 #include <plugins/gbp/gbp_recirc.h>
 
+#ifndef CLIB_MARCH_VARIANT
 /**
  * DPO pool
  */
@@ -206,6 +207,7 @@ gbp_policy_dpo_module_init (vlib_main_t * vm)
 }
 
 VLIB_INIT_FUNCTION (gbp_policy_dpo_module_init);
+#endif /* CLIB_MARCH_VARIANT */
 
 typedef struct gbp_policy_dpo_trace_t_
 {
@@ -394,23 +396,22 @@ format_gbp_policy_dpo_trace (u8 * s, va_list * args)
   return s;
 }
 
-static uword
-ip4_gbp_policy_dpo (vlib_main_t * vm,
-                   vlib_node_runtime_t * node, vlib_frame_t * from_frame)
+VLIB_NODE_FN (ip4_gbp_policy_dpo_node) (vlib_main_t * vm,
+                                       vlib_node_runtime_t * node,
+                                       vlib_frame_t * from_frame)
 {
   return (gbp_policy_dpo_inline (vm, node, from_frame, 0));
 }
 
-static uword
-ip6_gbp_policy_dpo (vlib_main_t * vm,
-                   vlib_node_runtime_t * node, vlib_frame_t * from_frame)
+VLIB_NODE_FN (ip6_gbp_policy_dpo_node) (vlib_main_t * vm,
+                                       vlib_node_runtime_t * node,
+                                       vlib_frame_t * from_frame)
 {
   return (gbp_policy_dpo_inline (vm, node, from_frame, 1));
 }
 
 /* *INDENT-OFF* */
 VLIB_REGISTER_NODE (ip4_gbp_policy_dpo_node) = {
-    .function = ip4_gbp_policy_dpo,
     .name = "ip4-gbp-policy-dpo",
     .vector_size = sizeof (u32),
     .format_trace = format_gbp_policy_dpo_trace,
@@ -421,7 +422,6 @@ VLIB_REGISTER_NODE (ip4_gbp_policy_dpo_node) = {
     }
 };
 VLIB_REGISTER_NODE (ip6_gbp_policy_dpo_node) = {
-    .function = ip6_gbp_policy_dpo,
     .name = "ip6-gbp-policy-dpo",
     .vector_size = sizeof (u32),
     .format_trace = format_gbp_policy_dpo_trace,
@@ -431,9 +431,6 @@ VLIB_REGISTER_NODE (ip6_gbp_policy_dpo_node) = {
         [GBP_POLICY_DROP] = "ip6-drop",
     }
 };
-
-VLIB_NODE_FUNCTION_MULTIARCH (ip4_gbp_policy_dpo_node, ip4_gbp_policy_dpo)
-VLIB_NODE_FUNCTION_MULTIARCH (ip6_gbp_policy_dpo_node, ip6_gbp_policy_dpo)
 /* *INDENT-ON* */
 
 /*
index e25ea38..24581e2 100644 (file)
@@ -146,44 +146,44 @@ gbp_sclass_inline (vlib_main_t * vm,
   return frame->n_vectors;
 }
 
-uword
-l2_gbp_id_2_sclass (vlib_main_t * vm,
-                   vlib_node_runtime_t * node, vlib_frame_t * frame)
+VLIB_NODE_FN (l2_gbp_id_2_sclass_node) (vlib_main_t * vm,
+                                       vlib_node_runtime_t * node,
+                                       vlib_frame_t * frame)
 {
   return (gbp_sclass_inline (vm, node, frame, 1, 1));
 }
 
-uword
-l2_gbp_sclass_2_id (vlib_main_t * vm,
-                   vlib_node_runtime_t * node, vlib_frame_t * frame)
+VLIB_NODE_FN (l2_gbp_sclass_2_id_node) (vlib_main_t * vm,
+                                       vlib_node_runtime_t * node,
+                                       vlib_frame_t * frame)
 {
   return (gbp_sclass_inline (vm, node, frame, 0, 1));
 }
 
-uword
-ip4_gbp_id_2_sclass (vlib_main_t * vm,
-                    vlib_node_runtime_t * node, vlib_frame_t * frame)
+VLIB_NODE_FN (ip4_gbp_id_2_sclass_node) (vlib_main_t * vm,
+                                        vlib_node_runtime_t * node,
+                                        vlib_frame_t * frame)
 {
   return (gbp_sclass_inline (vm, node, frame, 1, 0));
 }
 
-uword
-ip4_gbp_sclass_2_id (vlib_main_t * vm,
-                    vlib_node_runtime_t * node, vlib_frame_t * frame)
+VLIB_NODE_FN (ip4_gbp_sclass_2_id_node) (vlib_main_t * vm,
+                                        vlib_node_runtime_t * node,
+                                        vlib_frame_t * frame)
 {
   return (gbp_sclass_inline (vm, node, frame, 0, 0));
 }
 
-uword
-ip6_gbp_id_2_sclass (vlib_main_t * vm,
-                    vlib_node_runtime_t * node, vlib_frame_t * frame)
+VLIB_NODE_FN (ip6_gbp_id_2_sclass_node) (vlib_main_t * vm,
+                                        vlib_node_runtime_t * node,
+                                        vlib_frame_t * frame)
 {
   return (gbp_sclass_inline (vm, node, frame, 1, 0));
 }
 
-uword
-ip6_gbp_sclass_2_id (vlib_main_t * vm,
-                    vlib_node_runtime_t * node, vlib_frame_t * frame)
+VLIB_NODE_FN (ip6_gbp_sclass_2_id_node) (vlib_main_t * vm,
+                                        vlib_node_runtime_t * node,
+                                        vlib_frame_t * frame)
 {
   return (gbp_sclass_inline (vm, node, frame, 0, 0));
 }
@@ -203,7 +203,6 @@ format_gbp_sclass_trace (u8 * s, va_list * args)
 
 /* *INDENT-OFF* */
 VLIB_REGISTER_NODE (l2_gbp_id_2_sclass_node) = {
-  .function = l2_gbp_id_2_sclass,
   .name = "l2-gbp-id-2-sclass",
   .vector_size = sizeof (u32),
   .format_trace = format_gbp_sclass_trace,
@@ -216,7 +215,6 @@ VLIB_REGISTER_NODE (l2_gbp_id_2_sclass_node) = {
   },
 };
 VLIB_REGISTER_NODE (l2_gbp_sclass_2_id_node) = {
-  .function = l2_gbp_sclass_2_id,
   .name = "l2-gbp-sclass-2-id",
   .vector_size = sizeof (u32),
   .format_trace = format_gbp_sclass_trace,
@@ -230,7 +228,6 @@ VLIB_REGISTER_NODE (l2_gbp_sclass_2_id_node) = {
 };
 
 VLIB_REGISTER_NODE (ip4_gbp_id_2_sclass_node) = {
-  .function = ip4_gbp_id_2_sclass,
   .name = "ip4-gbp-id-2-sclass",
   .vector_size = sizeof (u32),
   .format_trace = format_gbp_sclass_trace,
@@ -243,7 +240,6 @@ VLIB_REGISTER_NODE (ip4_gbp_id_2_sclass_node) = {
   },
 };
 VLIB_REGISTER_NODE (ip4_gbp_sclass_2_id_node) = {
-  .function = ip4_gbp_sclass_2_id,
   .name = "ip4-gbp-sclass-2-id",
   .vector_size = sizeof (u32),
   .format_trace = format_gbp_sclass_trace,
@@ -257,7 +253,6 @@ VLIB_REGISTER_NODE (ip4_gbp_sclass_2_id_node) = {
 };
 
 VLIB_REGISTER_NODE (ip6_gbp_id_2_sclass_node) = {
-  .function = ip6_gbp_id_2_sclass,
   .name = "ip6-gbp-id-2-sclass",
   .vector_size = sizeof (u32),
   .format_trace = format_gbp_sclass_trace,
@@ -270,7 +265,6 @@ VLIB_REGISTER_NODE (ip6_gbp_id_2_sclass_node) = {
   },
 };
 VLIB_REGISTER_NODE (ip6_gbp_sclass_2_id_node) = {
-  .function = ip6_gbp_sclass_2_id,
   .name = "ip6-gbp-sclass-2-id",
   .vector_size = sizeof (u32),
   .format_trace = format_gbp_sclass_trace,
@@ -283,13 +277,6 @@ VLIB_REGISTER_NODE (ip6_gbp_sclass_2_id_node) = {
   },
 };
 
-VLIB_NODE_FUNCTION_MULTIARCH (l2_gbp_id_2_sclass_node, l2_gbp_id_2_sclass);
-VLIB_NODE_FUNCTION_MULTIARCH (l2_gbp_sclass_2_id_node, l2_gbp_sclass_2_id);
-
-VLIB_NODE_FUNCTION_MULTIARCH (ip4_gbp_id_2_sclass_node, ip4_gbp_id_2_sclass);
-VLIB_NODE_FUNCTION_MULTIARCH (ip4_gbp_sclass_2_id_node, ip4_gbp_sclass_2_id);
-VLIB_NODE_FUNCTION_MULTIARCH (ip6_gbp_id_2_sclass_node, ip6_gbp_id_2_sclass);
-VLIB_NODE_FUNCTION_MULTIARCH (ip6_gbp_sclass_2_id_node, ip6_gbp_sclass_2_id);
 
 VNET_FEATURE_INIT (ip4_gbp_sclass_2_id_feat, static) =
 {
@@ -315,6 +302,7 @@ VNET_FEATURE_INIT (ip6_gbp_id_2_sclass_feat, static) =
 };
 /* *INDENT-ON* */
 
+#ifndef CLIB_MARCH_VARIANT
 void
 gbp_sclass_enable_l2 (u32 sw_if_index)
 {
@@ -354,20 +342,24 @@ gbp_sclass_disable_ip (u32 sw_if_index)
   vnet_feature_enable_disable ("ip6-output",
                               "ip6-gbp-id-2-sclass", sw_if_index, 0, 0, 0);
 }
+#endif /* CLIB_MARCH_VARIANT */
 
 static clib_error_t *
 gbp_sclass_init (vlib_main_t * vm)
 {
   gbp_sclass_main_t *glm = &gbp_sclass_main;
+  vlib_node_t *node = vlib_get_node_by_name (vm, (u8 *) "l2-gbp-sclass-2-id");
 
   /* Initialize the feature next-node indices */
   feat_bitmap_init_next_nodes (vm,
-                              l2_gbp_sclass_2_id_node.index,
+                              node->index,
                               L2INPUT_N_FEAT,
                               l2input_get_feat_names (),
                               glm->gel_l2_input_feat_next);
+
+  node = vlib_get_node_by_name (vm, (u8 *) "l2-gbp-id-2-sclass");
   feat_bitmap_init_next_nodes (vm,
-                              l2_gbp_id_2_sclass_node.index,
+                              node->index,
                               L2OUTPUT_N_FEAT,
                               l2output_get_feat_names (),
                               glm->gel_l2_output_feat_next);
index c7be90f..846ea0f 100644 (file)
@@ -79,7 +79,7 @@ static char *gbp_vxlan_tunnel_layer_strings[] = {
 
 
 
-always_inline gbp_vxlan_tunnel_t *
+gbp_vxlan_tunnel_t *
 gbp_vxlan_tunnel_get (index_t gti)
 {
   return (pool_elt_at_index (gbp_vxlan_tunnel_pool, gti));
@@ -325,194 +325,6 @@ vxlan_gbp_tunnel_lock (u32 sw_if_index)
   vxr->vxr_locks++;
 }
 
-#define foreach_gbp_vxlan_input_next         \
-  _(DROP, "error-drop")                      \
-  _(L2_INPUT, "l2-input")                    \
-  _(IP4_INPUT, "ip4-input")                  \
-  _(IP6_INPUT, "ip6-input")
-
-typedef enum
-{
-#define _(s,n) GBP_VXLAN_INPUT_NEXT_##s,
-  foreach_gbp_vxlan_input_next
-#undef _
-    GBP_VXLAN_INPUT_N_NEXT,
-} gbp_vxlan_input_next_t;
-
-#define foreach_gbp_vxlan_error              \
-  _(DECAPPED, "decapped")                    \
-  _(LEARNED, "learned")
-
-typedef enum
-{
-#define _(s,n) GBP_VXLAN_ERROR_##s,
-  foreach_gbp_vxlan_error
-#undef _
-    GBP_VXLAN_N_ERROR,
-} gbp_vxlan_input_error_t;
-
-static char *gbp_vxlan_error_strings[] = {
-#define _(n,s) s,
-  foreach_gbp_vxlan_error
-#undef _
-};
-
-typedef struct gbp_vxlan_trace_t_
-{
-  u8 dropped;
-  u32 vni;
-  u32 sw_if_index;
-  u16 sclass;
-  u8 flags;
-} gbp_vxlan_trace_t;
-
-
-static uword
-gbp_vxlan_decap (vlib_main_t * vm,
-                vlib_node_runtime_t * node,
-                vlib_frame_t * from_frame, u8 is_ip4)
-{
-  u32 n_left_to_next, n_left_from, next_index, *to_next, *from;
-
-  next_index = 0;
-  from = vlib_frame_vector_args (from_frame);
-  n_left_from = from_frame->n_vectors;
-
-  while (n_left_from > 0)
-    {
-
-      vlib_get_next_frame (vm, node, next_index, to_next, n_left_to_next);
-
-      while (n_left_from > 0 && n_left_to_next > 0)
-       {
-         vxlan_gbp_header_t *vxlan_gbp0;
-         gbp_vxlan_input_next_t next0;
-         gbp_vxlan_tunnel_t *gt0;
-         vlib_buffer_t *b0;
-         u32 bi0, vni0;
-         uword *p;
-
-         bi0 = to_next[0] = from[0];
-         from += 1;
-         to_next += 1;
-         n_left_from -= 1;
-         n_left_to_next -= 1;
-         next0 = GBP_VXLAN_INPUT_NEXT_DROP;
-
-         b0 = vlib_get_buffer (vm, bi0);
-         vxlan_gbp0 =
-           vlib_buffer_get_current (b0) - sizeof (vxlan_gbp_header_t);
-
-         vni0 = vxlan_gbp_get_vni (vxlan_gbp0);
-         p = hash_get (gv_db, vni0);
-
-         if (PREDICT_FALSE (NULL == p))
-           {
-             gt0 = NULL;
-             next0 = GBP_VXLAN_INPUT_NEXT_DROP;
-           }
-         else
-           {
-             gt0 = gbp_vxlan_tunnel_get (p[0]);
-
-             vnet_buffer (b0)->sw_if_index[VLIB_RX] = gt0->gt_sw_if_index;
-
-             if (GBP_VXLAN_TUN_L2 == gt0->gt_layer)
-               /*
-                * An L2 layer tunnel goes into the BD
-                */
-               next0 = GBP_VXLAN_INPUT_NEXT_L2_INPUT;
-             else
-               {
-                 /*
-                  * An L3 layer tunnel needs to strip the L2 header
-                  * an inject into the RD
-                  */
-                 ethernet_header_t *e0;
-                 u16 type0;
-
-                 e0 = vlib_buffer_get_current (b0);
-                 type0 = clib_net_to_host_u16 (e0->type);
-                 switch (type0)
-                   {
-                   case ETHERNET_TYPE_IP4:
-                     next0 = GBP_VXLAN_INPUT_NEXT_IP4_INPUT;
-                     break;
-                   case ETHERNET_TYPE_IP6:
-                     next0 = GBP_VXLAN_INPUT_NEXT_IP6_INPUT;
-                     break;
-                   default:
-                     goto trace;
-                   }
-                 vlib_buffer_advance (b0, sizeof (*e0));
-               }
-           }
-
-       trace:
-         if (PREDICT_FALSE (b0->flags & VLIB_BUFFER_IS_TRACED))
-           {
-             gbp_vxlan_trace_t *tr;
-
-             tr = vlib_add_trace (vm, node, b0, sizeof (*tr));
-             tr->dropped = (next0 == GBP_VXLAN_INPUT_NEXT_DROP);
-             tr->vni = vni0;
-             tr->sw_if_index = (gt0 ? gt0->gt_sw_if_index : ~0);
-             tr->flags = vxlan_gbp_get_gpflags (vxlan_gbp0);
-             tr->sclass = vxlan_gbp_get_sclass (vxlan_gbp0);
-           }
-
-         vlib_validate_buffer_enqueue_x1 (vm, node, next_index,
-                                          to_next, n_left_to_next,
-                                          bi0, next0);
-       }
-
-      vlib_put_next_frame (vm, node, next_index, n_left_to_next);
-    }
-
-  return from_frame->n_vectors;
-}
-
-static u8 *
-format_gbp_vxlan_rx_trace (u8 * s, va_list * args)
-{
-  CLIB_UNUSED (vlib_main_t * vm) = va_arg (*args, vlib_main_t *);
-  CLIB_UNUSED (vlib_node_t * node) = va_arg (*args, vlib_node_t *);
-  gbp_vxlan_trace_t *t = va_arg (*args, gbp_vxlan_trace_t *);
-
-  s = format (s, "vni:%d dropped:%d rx:%d sclass:%d flags:%U",
-             t->vni, t->dropped, t->sw_if_index,
-             t->sclass, format_vxlan_gbp_header_gpflags, t->flags);
-
-  return (s);
-}
-
-static uword
-gbp_vxlan4_decap (vlib_main_t * vm,
-                 vlib_node_runtime_t * node, vlib_frame_t * from_frame)
-{
-  return gbp_vxlan_decap (vm, node, from_frame, 1);
-}
-
-/* *INDENT-OFF* */
-VLIB_REGISTER_NODE (gbp_vxlan4_input_node) =
-{
-  .function = gbp_vxlan4_decap,
-  .name = "gbp-vxlan4",
-  .vector_size = sizeof (u32),
-  .n_errors = GBP_VXLAN_N_ERROR,
-  .error_strings = gbp_vxlan_error_strings,
-  .n_next_nodes = GBP_VXLAN_INPUT_N_NEXT,
-  .format_trace = format_gbp_vxlan_rx_trace,
-  .next_nodes = {
-#define _(s,n) [GBP_VXLAN_INPUT_NEXT_##s] = n,
-    foreach_gbp_vxlan_input_next
-#undef _
-  },
-};
-VLIB_NODE_FUNCTION_MULTIARCH (gbp_vxlan4_input_node, gbp_vxlan4_decap)
-
-/* *INDENT-ON* */
-
 void
 gbp_vxlan_walk (gbp_vxlan_cb_t cb, void *ctx)
 {
@@ -866,6 +678,7 @@ static clib_error_t *
 gbp_vxlan_init (vlib_main_t * vm)
 {
   u32 slot4;
+  vlib_node_t *node = vlib_get_node_by_name (vm, (u8 *) "gbp-vxlan4");
 
   /*
    * insert ourselves into the VXLAN-GBP arc to collect the no-tunnel
@@ -873,7 +686,7 @@ gbp_vxlan_init (vlib_main_t * vm)
    */
   slot4 = vlib_node_add_next_with_slot (vm,
                                        vxlan4_gbp_input_node.index,
-                                       gbp_vxlan4_input_node.index,
+                                       node->index,
                                        VXLAN_GBP_INPUT_NEXT_NO_TUNNEL);
   ASSERT (slot4 == VXLAN_GBP_INPUT_NEXT_NO_TUNNEL);
 
index 7aa22e3..acf5f4b 100644 (file)
@@ -122,6 +122,7 @@ extern void gbp_vxlan_walk (gbp_vxlan_cb_t cb, void *ctx);
 extern u8 *format_gbp_vxlan_tunnel (u8 * s, va_list * args);
 extern u8 *format_gbp_vxlan_tunnel_layer (u8 * s, va_list * args);
 
+extern gbp_vxlan_tunnel_t *gbp_vxlan_tunnel_get (index_t gti);
 #endif
 
 /*
diff --git a/src/plugins/gbp/gbp_vxlan_node.c b/src/plugins/gbp/gbp_vxlan_node.c
new file mode 100644 (file)
index 0000000..c49f74a
--- /dev/null
@@ -0,0 +1,219 @@
+/*
+ * Copyright (c) 2019 Cisco and/or its affiliates.
+ * Licensed under the Apache License, Version 2.0 (the "License");
+ * you may not use this file except in compliance with the License.
+ * You may obtain a copy of the License at:
+ *
+ *     http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+#include <plugins/gbp/gbp_vxlan.h>
+#include <plugins/gbp/gbp_itf.h>
+#include <plugins/gbp/gbp_learn.h>
+#include <plugins/gbp/gbp_bridge_domain.h>
+#include <plugins/gbp/gbp_route_domain.h>
+#include <plugins/gbp/gbp_sclass.h>
+
+#include <vnet/vxlan-gbp/vxlan_gbp.h>
+#include <vlibmemory/api.h>
+#include <vnet/fib/fib_table.h>
+
+extern uword *gv_db;
+
+typedef struct gbp_vxlan_trace_t_
+{
+  u8 dropped;
+  u32 vni;
+  u32 sw_if_index;
+  u16 sclass;
+  u8 flags;
+} gbp_vxlan_trace_t;
+
+#define foreach_gbp_vxlan_input_next         \
+  _(DROP, "error-drop")                      \
+  _(L2_INPUT, "l2-input")                    \
+  _(IP4_INPUT, "ip4-input")                  \
+  _(IP6_INPUT, "ip6-input")
+
+typedef enum
+{
+#define _(s,n) GBP_VXLAN_INPUT_NEXT_##s,
+  foreach_gbp_vxlan_input_next
+#undef _
+    GBP_VXLAN_INPUT_N_NEXT,
+} gbp_vxlan_input_next_t;
+
+
+#define foreach_gbp_vxlan_error              \
+  _(DECAPPED, "decapped")                    \
+  _(LEARNED, "learned")
+
+typedef enum
+{
+#define _(s,n) GBP_VXLAN_ERROR_##s,
+  foreach_gbp_vxlan_error
+#undef _
+    GBP_VXLAN_N_ERROR,
+} gbp_vxlan_input_error_t;
+
+static char *gbp_vxlan_error_strings[] = {
+#define _(n,s) s,
+  foreach_gbp_vxlan_error
+#undef _
+};
+
+static uword
+gbp_vxlan_decap (vlib_main_t * vm,
+                vlib_node_runtime_t * node,
+                vlib_frame_t * from_frame, u8 is_ip4)
+{
+  u32 n_left_to_next, n_left_from, next_index, *to_next, *from;
+
+  next_index = 0;
+  from = vlib_frame_vector_args (from_frame);
+  n_left_from = from_frame->n_vectors;
+
+  while (n_left_from > 0)
+    {
+
+      vlib_get_next_frame (vm, node, next_index, to_next, n_left_to_next);
+
+      while (n_left_from > 0 && n_left_to_next > 0)
+       {
+         vxlan_gbp_header_t *vxlan_gbp0;
+         gbp_vxlan_input_next_t next0;
+         gbp_vxlan_tunnel_t *gt0;
+         vlib_buffer_t *b0;
+         u32 bi0, vni0;
+         uword *p;
+
+         bi0 = to_next[0] = from[0];
+         from += 1;
+         to_next += 1;
+         n_left_from -= 1;
+         n_left_to_next -= 1;
+         next0 = GBP_VXLAN_INPUT_NEXT_DROP;
+
+         b0 = vlib_get_buffer (vm, bi0);
+         vxlan_gbp0 =
+           vlib_buffer_get_current (b0) - sizeof (vxlan_gbp_header_t);
+
+         vni0 = vxlan_gbp_get_vni (vxlan_gbp0);
+         p = hash_get (gv_db, vni0);
+
+         if (PREDICT_FALSE (NULL == p))
+           {
+             gt0 = NULL;
+             next0 = GBP_VXLAN_INPUT_NEXT_DROP;
+           }
+         else
+           {
+             gt0 = gbp_vxlan_tunnel_get (p[0]);
+
+             vnet_buffer (b0)->sw_if_index[VLIB_RX] = gt0->gt_sw_if_index;
+
+             if (GBP_VXLAN_TUN_L2 == gt0->gt_layer)
+               /*
+                * An L2 layer tunnel goes into the BD
+                */
+               next0 = GBP_VXLAN_INPUT_NEXT_L2_INPUT;
+             else
+               {
+                 /*
+                  * An L3 layer tunnel needs to strip the L2 header
+                  * an inject into the RD
+                  */
+                 ethernet_header_t *e0;
+                 u16 type0;
+
+                 e0 = vlib_buffer_get_current (b0);
+                 type0 = clib_net_to_host_u16 (e0->type);
+                 switch (type0)
+                   {
+                   case ETHERNET_TYPE_IP4:
+                     next0 = GBP_VXLAN_INPUT_NEXT_IP4_INPUT;
+                     break;
+                   case ETHERNET_TYPE_IP6:
+                     next0 = GBP_VXLAN_INPUT_NEXT_IP6_INPUT;
+                     break;
+                   default:
+                     goto trace;
+                   }
+                 vlib_buffer_advance (b0, sizeof (*e0));
+               }
+           }
+
+       trace:
+         if (PREDICT_FALSE (b0->flags & VLIB_BUFFER_IS_TRACED))
+           {
+             gbp_vxlan_trace_t *tr;
+
+             tr = vlib_add_trace (vm, node, b0, sizeof (*tr));
+             tr->dropped = (next0 == GBP_VXLAN_INPUT_NEXT_DROP);
+             tr->vni = vni0;
+             tr->sw_if_index = (gt0 ? gt0->gt_sw_if_index : ~0);
+             tr->flags = vxlan_gbp_get_gpflags (vxlan_gbp0);
+             tr->sclass = vxlan_gbp_get_sclass (vxlan_gbp0);
+           }
+
+         vlib_validate_buffer_enqueue_x1 (vm, node, next_index,
+                                          to_next, n_left_to_next,
+                                          bi0, next0);
+       }
+
+      vlib_put_next_frame (vm, node, next_index, n_left_to_next);
+    }
+
+  return from_frame->n_vectors;
+}
+
+VLIB_NODE_FN (gbp_vxlan4_input_node) (vlib_main_t * vm,
+                                     vlib_node_runtime_t * node,
+                                     vlib_frame_t * from_frame)
+{
+  return gbp_vxlan_decap (vm, node, from_frame, 1);
+}
+
+static u8 *
+format_gbp_vxlan_rx_trace (u8 * s, va_list * args)
+{
+  CLIB_UNUSED (vlib_main_t * vm) = va_arg (*args, vlib_main_t *);
+  CLIB_UNUSED (vlib_node_t * node) = va_arg (*args, vlib_node_t *);
+  gbp_vxlan_trace_t *t = va_arg (*args, gbp_vxlan_trace_t *);
+
+  s = format (s, "vni:%d dropped:%d rx:%d sclass:%d flags:%U",
+             t->vni, t->dropped, t->sw_if_index,
+             t->sclass, format_vxlan_gbp_header_gpflags, t->flags);
+
+  return (s);
+}
+
+/* *INDENT-OFF* */
+VLIB_REGISTER_NODE (gbp_vxlan4_input_node) =
+{
+  .name = "gbp-vxlan4",
+  .vector_size = sizeof (u32),
+  .n_errors = GBP_VXLAN_N_ERROR,
+  .error_strings = gbp_vxlan_error_strings,
+  .n_next_nodes = GBP_VXLAN_INPUT_N_NEXT,
+  .format_trace = format_gbp_vxlan_rx_trace,
+  .next_nodes = {
+#define _(s,n) [GBP_VXLAN_INPUT_NEXT_##s] = n,
+    foreach_gbp_vxlan_input_next
+#undef _
+  },
+};
+/* *INDENT-ON* */
+
+/*
+ * fd.io coding-style-patch-verification: ON
+ *
+ * Local Variables:
+ * eval: (c-set-style "gnu")
+ * End:
+ */