X-Git-Url: https://gerrit.fd.io/r/gitweb?a=blobdiff_plain;f=src%2Fplugins%2Fgbp%2Fgbp_vxlan.c;h=286345ed8fa562751c2a086dbe0a309a8fdb43c7;hb=4ba67723d;hp=b29fc111131e9374a40099ed9fb1b65c9979640d;hpb=93cc3ee3b3a9c9224a1446625882205f3282a949;p=vpp.git diff --git a/src/plugins/gbp/gbp_vxlan.c b/src/plugins/gbp/gbp_vxlan.c index b29fc111131..286345ed8fa 100644 --- a/src/plugins/gbp/gbp_vxlan.c +++ b/src/plugins/gbp/gbp_vxlan.c @@ -78,7 +78,7 @@ static char *gbp_vxlan_tunnel_layer_strings[] = { -always_inline gbp_vxlan_tunnel_t * +gbp_vxlan_tunnel_t * gbp_vxlan_tunnel_get (index_t gti) { return (pool_elt_at_index (gbp_vxlan_tunnel_pool, gti)); @@ -106,13 +106,12 @@ format_vxlan_tunnel_ref (u8 * s, va_list * args) static u32 gdb_vxlan_dep_add (gbp_vxlan_tunnel_t * gt, - u32 vni, const ip46_address_t * src, const ip46_address_t * dst) { vnet_vxlan_gbp_tunnel_add_del_args_t args = { .is_add = 1, .is_ip6 = !ip46_address_is_ip4 (src), - .vni = vni, + .vni = gt->gt_vni, .src = *src, .dst = *dst, .instance = ~0, @@ -140,7 +139,7 @@ gdb_vxlan_dep_add (gbp_vxlan_tunnel_t * gt, GBP_VXLAN_TUN_DBG ("add-dep:%U %U %U %d", format_vnet_sw_if_index_name, vnet_get_main (), sw_if_index, format_ip46_address, src, IP46_TYPE_ANY, - format_ip46_address, dst, IP46_TYPE_ANY, vni); + format_ip46_address, dst, IP46_TYPE_ANY, gt->gt_vni); pool_get_zero (vxlan_tunnel_ref_pool, vxr); @@ -161,13 +160,24 @@ gdb_vxlan_dep_add (gbp_vxlan_tunnel_t * gt, if (GBP_VXLAN_TUN_L2 == vxr->vxr_layer) { + l2output_feat_masks_t ofeat; + l2input_feat_masks_t ifeat; + gbp_bridge_domain_t *gbd; + + gbd = gbp_bridge_domain_get (gt->gt_gbd); vxr->vxr_itf = gbp_itf_add_and_lock (vxr->vxr_sw_if_index, gt->gt_bd_index); - gbp_itf_set_l2_output_feature (vxr->vxr_itf, vxr->vxr_sw_if_index, - L2OUTPUT_FEAT_GBP_POLICY_MAC); - gbp_itf_set_l2_input_feature (vxr->vxr_itf, vxr->vxr_sw_if_index, - L2INPUT_FEAT_GBP_LEARN); + ofeat = L2OUTPUT_FEAT_GBP_POLICY_MAC; + ifeat = L2INPUT_FEAT_NONE; + + if (!(gbd->gb_flags & GBP_BD_FLAG_DO_NOT_LEARN)) + ifeat |= L2INPUT_FEAT_GBP_LEARN; + + gbp_itf_set_l2_output_feature (vxr->vxr_itf, + vxr->vxr_sw_if_index, ofeat); + gbp_itf_set_l2_input_feature (vxr->vxr_itf, + vxr->vxr_sw_if_index, ifeat); } else { @@ -235,7 +245,7 @@ gbp_vxlan_tunnel_clone_and_lock (u32 sw_if_index, gt = pool_elt_at_index (gbp_vxlan_tunnel_pool, gti); - return (gdb_vxlan_dep_add (gt, gt->gt_vni, src, dst)); + return (gdb_vxlan_dep_add (gt, src, dst)); } static void @@ -270,6 +280,7 @@ gdb_vxlan_dep_del (index_t vxri) FOR_EACH_FIB_IP_PROTOCOL (fproto) ip_table_bind (fproto, vxr->vxr_sw_if_index, 0, 0); + gbp_learn_disable (vxr->vxr_sw_if_index, GBP_LEARN_MODE_L3); } vnet_vxlan_gbp_tunnel_del (vxr->vxr_sw_if_index); @@ -310,194 +321,6 @@ vxlan_gbp_tunnel_lock (u32 sw_if_index) vxr->vxr_locks++; } -#define foreach_gbp_vxlan_input_next \ - _(DROP, "error-drop") \ - _(L2_INPUT, "l2-input") \ - _(IP4_INPUT, "ip4-input") \ - _(IP6_INPUT, "ip6-input") - -typedef enum -{ -#define _(s,n) GBP_VXLAN_INPUT_NEXT_##s, - foreach_gbp_vxlan_input_next -#undef _ - GBP_VXLAN_INPUT_N_NEXT, -} gbp_vxlan_input_next_t; - -#define foreach_gbp_vxlan_error \ - _(DECAPPED, "decapped") \ - _(LEARNED, "learned") - -typedef enum -{ -#define _(s,n) GBP_VXLAN_ERROR_##s, - foreach_gbp_vxlan_error -#undef _ - GBP_VXLAN_N_ERROR, -} gbp_vxlan_input_error_t; - -static char *gbp_vxlan_error_strings[] = { -#define _(n,s) s - foreach_gbp_vxlan_error -#undef _ -}; - -typedef struct gbp_vxlan_trace_t_ -{ - u8 dropped; - u32 vni; - u32 sw_if_index; - u16 sclass; - u8 flags; -} gbp_vxlan_trace_t; - - -static uword -gbp_vxlan_decap (vlib_main_t * vm, - vlib_node_runtime_t * node, - vlib_frame_t * from_frame, u8 is_ip4) -{ - u32 n_left_to_next, n_left_from, next_index, *to_next, *from; - - next_index = 0; - from = vlib_frame_vector_args (from_frame); - n_left_from = from_frame->n_vectors; - - while (n_left_from > 0) - { - - vlib_get_next_frame (vm, node, next_index, to_next, n_left_to_next); - - while (n_left_from > 0 && n_left_to_next > 0) - { - vxlan_gbp_header_t *vxlan_gbp0; - gbp_vxlan_input_next_t next0; - gbp_vxlan_tunnel_t *gt0; - vlib_buffer_t *b0; - u32 bi0, vni0; - uword *p; - - bi0 = to_next[0] = from[0]; - from += 1; - to_next += 1; - n_left_from -= 1; - n_left_to_next -= 1; - next0 = GBP_VXLAN_INPUT_NEXT_DROP; - - b0 = vlib_get_buffer (vm, bi0); - vxlan_gbp0 = - vlib_buffer_get_current (b0) - sizeof (vxlan_gbp_header_t); - - vni0 = vxlan_gbp_get_vni (vxlan_gbp0); - p = hash_get (gv_db, vni0); - - if (PREDICT_FALSE (NULL == p)) - { - gt0 = NULL; - next0 = GBP_VXLAN_INPUT_NEXT_DROP; - } - else - { - gt0 = gbp_vxlan_tunnel_get (p[0]); - - vnet_buffer (b0)->sw_if_index[VLIB_RX] = gt0->gt_sw_if_index; - - if (GBP_VXLAN_TUN_L2 == gt0->gt_layer) - /* - * An L2 layer tunnel goes into the BD - */ - next0 = GBP_VXLAN_INPUT_NEXT_L2_INPUT; - else - { - /* - * An L3 layer tunnel needs to strip the L2 header - * an inject into the RD - */ - ethernet_header_t *e0; - u16 type0; - - e0 = vlib_buffer_get_current (b0); - type0 = clib_net_to_host_u16 (e0->type); - switch (type0) - { - case ETHERNET_TYPE_IP4: - next0 = GBP_VXLAN_INPUT_NEXT_IP4_INPUT; - break; - case ETHERNET_TYPE_IP6: - next0 = GBP_VXLAN_INPUT_NEXT_IP6_INPUT; - break; - default: - goto trace; - } - vlib_buffer_advance (b0, sizeof (*e0)); - } - } - - trace: - if (PREDICT_FALSE (b0->flags & VLIB_BUFFER_IS_TRACED)) - { - gbp_vxlan_trace_t *tr; - - tr = vlib_add_trace (vm, node, b0, sizeof (*tr)); - tr->dropped = (next0 == GBP_VXLAN_INPUT_NEXT_DROP); - tr->vni = vni0; - tr->sw_if_index = (gt0 ? gt0->gt_sw_if_index : ~0); - tr->flags = vxlan_gbp_get_gpflags (vxlan_gbp0); - tr->sclass = vxlan_gbp_get_sclass (vxlan_gbp0); - } - - vlib_validate_buffer_enqueue_x1 (vm, node, next_index, - to_next, n_left_to_next, - bi0, next0); - } - - vlib_put_next_frame (vm, node, next_index, n_left_to_next); - } - - return from_frame->n_vectors; -} - -static u8 * -format_gbp_vxlan_rx_trace (u8 * s, va_list * args) -{ - CLIB_UNUSED (vlib_main_t * vm) = va_arg (*args, vlib_main_t *); - CLIB_UNUSED (vlib_node_t * node) = va_arg (*args, vlib_node_t *); - gbp_vxlan_trace_t *t = va_arg (*args, gbp_vxlan_trace_t *); - - s = format (s, "vni:%d dropped:%d rx:%d sclass:%d flags:%U", - t->vni, t->dropped, t->sw_if_index, - t->sclass, format_vxlan_gbp_header_gpflags, t->flags); - - return (s); -} - -static uword -gbp_vxlan4_decap (vlib_main_t * vm, - vlib_node_runtime_t * node, vlib_frame_t * from_frame) -{ - return gbp_vxlan_decap (vm, node, from_frame, 1); -} - -/* *INDENT-OFF* */ -VLIB_REGISTER_NODE (gbp_vxlan4_input_node) = -{ - .function = gbp_vxlan4_decap, - .name = "gbp-vxlan4", - .vector_size = sizeof (u32), - .n_errors = GBP_VXLAN_N_ERROR, - .error_strings = gbp_vxlan_error_strings, - .n_next_nodes = GBP_VXLAN_INPUT_N_NEXT, - .format_trace = format_gbp_vxlan_rx_trace, - .next_nodes = { -#define _(s,n) [GBP_VXLAN_INPUT_NEXT_##s] = n, - foreach_gbp_vxlan_input_next -#undef _ - }, -}; -VLIB_NODE_FUNCTION_MULTIARCH (gbp_vxlan4_input_node, gbp_vxlan4_decap) - -/* *INDENT-ON* */ - void gbp_vxlan_walk (gbp_vxlan_cb_t cb, void *ctx) { @@ -638,7 +461,8 @@ VNET_HW_INTERFACE_CLASS (gbp_vxlan_hw_interface_class) = { int gbp_vxlan_tunnel_add (u32 vni, gbp_vxlan_tunnel_layer_t layer, - u32 bd_rd_id, u32 * sw_if_indexp) + u32 bd_rd_id, + const ip4_address_t * src, u32 * sw_if_indexp) { gbp_vxlan_tunnel_t *gt; index_t gti; @@ -685,6 +509,7 @@ gbp_vxlan_tunnel_add (u32 vni, gbp_vxlan_tunnel_layer_t layer, gt->gt_vni = vni; gt->gt_layer = layer; gt->gt_bd_rd_id = bd_rd_id; + gt->gt_src.ip4.as_u32 = src->as_u32; gt->gt_hw_if_index = vnet_register_interface (vnm, gbp_vxlan_device_class.index, gti, @@ -706,8 +531,8 @@ gbp_vxlan_tunnel_add (u32 vni, gbp_vxlan_tunnel_layer_t layer, gb = gbp_bridge_domain_get (gbi); gt->gt_gbd = gbi; - gt->gt_bd_index = gb->gb_bd_id; - gb->gb_vni_sw_if_index = gt->gt_sw_if_index; + gt->gt_bd_index = gb->gb_bd_index; + gb->gb_vni = gti; /* set it up as a GBP interface */ gt->gt_itf = gbp_itf_add_and_lock (gt->gt_sw_if_index, gt->gt_bd_index); @@ -742,7 +567,8 @@ gbp_vxlan_tunnel_add (u32 vni, gbp_vxlan_tunnel_layer_t layer, */ hash_set (gv_db, vni, gti); - vec_validate (gbp_vxlan_tunnel_db, gt->gt_sw_if_index); + vec_validate_init_empty (gbp_vxlan_tunnel_db, + gt->gt_sw_if_index, INDEX_INVALID); gbp_vxlan_tunnel_db[gt->gt_sw_if_index] = gti; if (sw_if_indexp) @@ -781,7 +607,7 @@ gbp_vxlan_tunnel_del (u32 vni) GBP_VXLAN_TUN_DBG ("del: %U", format_gbp_vxlan_tunnel, gt - gbp_vxlan_tunnel_pool); - gbp_endpoint_flush (gt->gt_sw_if_index); + gbp_endpoint_flush (GBP_ENDPOINT_SRC_DP, gt->gt_sw_if_index); ASSERT (0 == vec_len (gt->gt_tuns)); vec_free (gt->gt_tuns); @@ -847,6 +673,7 @@ static clib_error_t * gbp_vxlan_init (vlib_main_t * vm) { u32 slot4; + vlib_node_t *node = vlib_get_node_by_name (vm, (u8 *) "gbp-vxlan4"); /* * insert ourselves into the VXLAN-GBP arc to collect the no-tunnel @@ -854,7 +681,7 @@ gbp_vxlan_init (vlib_main_t * vm) */ slot4 = vlib_node_add_next_with_slot (vm, vxlan4_gbp_input_node.index, - gbp_vxlan4_input_node.index, + node->index, VXLAN_GBP_INPUT_NEXT_NO_TUNNEL); ASSERT (slot4 == VXLAN_GBP_INPUT_NEXT_NO_TUNNEL);