X-Git-Url: https://gerrit.fd.io/r/gitweb?a=blobdiff_plain;f=src%2Fvnet%2Fl2%2Fl2_input.c;h=fbf4a6b499a44a4879633087fc4252239cdf3a5e;hb=b1232555e91b286feab5667b5a22f29aa8e96626;hp=aca23fe03fcfcb29689a85520d6a5d661bece3b9;hpb=9793477a28c45e4eb5bba3f2050fe415e57e8ad8;p=vpp.git diff --git a/src/vnet/l2/l2_input.c b/src/vnet/l2/l2_input.c index aca23fe03fc..fbf4a6b499a 100644 --- a/src/vnet/l2/l2_input.c +++ b/src/vnet/l2/l2_input.c @@ -23,6 +23,8 @@ #include #include #include +#include +#include #include #include #include @@ -60,6 +62,29 @@ l2input_get_feat_names (void) return l2input_feat_names; } +u8 * +format_l2_input_features (u8 * s, va_list * args) +{ + static char *display_names[] = { +#define _(sym,name) #sym, + foreach_l2input_feat +#undef _ + }; + u32 feature_bitmap = va_arg (*args, u32); + + if (feature_bitmap == 0) + { + s = format (s, " none configured"); + return s; + } + + feature_bitmap &= ~L2INPUT_FEAT_DROP; /* Not a feature */ + int i; + for (i = L2INPUT_N_FEAT; i >= 0; i--) + if (feature_bitmap & (1 << i)) + s = format (s, "%10s (%s)\n", display_names[i], l2input_feat_names[i]); + return s; +} typedef struct { @@ -113,12 +138,85 @@ typedef enum L2INPUT_N_NEXT, } l2input_next_t; +static_always_inline u32 +l2_input_classiy_unicast (vlib_buffer_t * b, const ethernet_header_t * h0) +{ + u32 feat_mask = ~0; + u8 *l3h0 = (u8 *) h0 + vnet_buffer (b)->l2.l2_len; + +#define get_u16(addr) ( *((u16 *)(addr)) ) + u16 ethertype = clib_net_to_host_u16 (get_u16 (l3h0 - 2)); + u8 protocol = ((ip6_header_t *) l3h0)->protocol; + + /* Disable bridge forwarding (flooding will execute instead if not xconnect) */ + feat_mask &= ~(L2INPUT_FEAT_FWD | + L2INPUT_FEAT_UU_FLOOD | L2INPUT_FEAT_GBP_FWD); + + /* Disable ARP-term for non-ARP and non-ICMP6 packet */ + if (ethertype != ETHERNET_TYPE_ARP && + (ethertype != ETHERNET_TYPE_IP6 || protocol != IP_PROTOCOL_ICMP6)) + feat_mask &= ~(L2INPUT_FEAT_ARP_TERM); + + /* + * For packet from BVI - set SHG of ARP request or ICMPv6 neighbor + * solicitation packet from BVI to 0 so it can also flood to VXLAN + * tunnels or other ports with the same SHG as that of the BVI. + */ + else if (PREDICT_FALSE (vnet_buffer (b)->sw_if_index[VLIB_TX] == + L2INPUT_BVI)) + { + if (ethertype == ETHERNET_TYPE_ARP) + { + ethernet_arp_header_t *arp0 = (ethernet_arp_header_t *) l3h0; + if (arp0->opcode == + clib_host_to_net_u16 (ETHERNET_ARP_OPCODE_request)) + vnet_buffer (b)->l2.shg = 0; + } + else /* must be ICMPv6 */ + { + ip6_header_t *iph0 = (ip6_header_t *) l3h0; + icmp6_neighbor_solicitation_or_advertisement_header_t *ndh0; + ndh0 = ip6_next_header (iph0); + if (ndh0->icmp.type == ICMP6_neighbor_solicitation) + vnet_buffer (b)->l2.shg = 0; + } + } + + return (feat_mask); +} static_always_inline void -classify_and_dispatch (vlib_main_t * vm, - vlib_node_runtime_t * node, - u32 thread_index, - l2input_main_t * msm, vlib_buffer_t * b0, u32 * next0) +l2_input_classify_bridge (l2_bridge_domain_t * bd_config, + u16 bd_index, + u32 sw_if_index, vlib_buffer_t * b, u32 * feat_mask) +{ + /* save BD ID for next feature graph nodes */ + vnet_buffer (b)->l2.bd_index = bd_index; + + /* Save bridge domain and interface seq_num */ + /* *INDENT-OFF* */ + l2fib_seq_num_t sn = { + .swif = *l2fib_swif_seq_num(sw_if_index), + .bd = bd_config->seq_num, + }; + /* *INDENT-ON* */ + vnet_buffer (b)->l2.l2fib_sn = sn.as_u16;; + vnet_buffer (b)->l2.bd_age = bd_config->mac_age; + + /* + * Process bridge domain feature enables. + * To perform learning/flooding/forwarding, the corresponding bit + * must be enabled in both the input interface config and in the + * bridge domain config. In the bd_bitmap, bits for features other + * than learning/flooding/forwarding should always be set. + */ + *feat_mask = *feat_mask & bd_config->feature_bitmap; +} + +static_always_inline void +classify_and_dispatch (l2input_main_t * msm, + u16 n_left, + u32 sw_if_index, vlib_buffer_t ** b, u16 * next) { /* * Load L2 input feature struct @@ -134,102 +232,202 @@ classify_and_dispatch (vlib_main_t * vm, * set tx sw-if-handle */ - u16 ethertype; - u8 protocol; l2_input_config_t *config; - l2_bridge_domain_t *bd_config; - u16 bd_index0; - u32 feature_bitmap; - u32 feat_mask; - ethernet_header_t *h0; - u8 *l3h0; - u32 sw_if_index0; -#define get_u16(addr) ( *((u16 *)(addr)) ) - - sw_if_index0 = vnet_buffer (b0)->sw_if_index[VLIB_RX]; + /* Get config for the input interface */ + config = vec_elt_at_index (msm->configs, sw_if_index); - h0 = vlib_buffer_get_current (b0); - l3h0 = (u8 *) h0 + vnet_buffer (b0)->l2.l2_len; + while (n_left > 8) + { + const ethernet_header_t *h0, *h1, *h2, *h3; + u32 fm0, fm1, fm2, fm3; + u32 fb0, fb1, fb2, fb3; - ethertype = clib_net_to_host_u16 (get_u16 (l3h0 - 2)); - feat_mask = ~0; + if (n_left >= 4) + { + vlib_prefetch_buffer_header (b[0], LOAD); + vlib_prefetch_buffer_header (b[1], LOAD); + vlib_prefetch_buffer_header (b[2], LOAD); + vlib_prefetch_buffer_header (b[3], LOAD); + } - /* Get config for the input interface */ - config = vec_elt_at_index (msm->configs, sw_if_index0); + fm0 = fm1 = fm2 = fm3 = ~0; + h0 = vlib_buffer_get_current (b[0]); + h1 = vlib_buffer_get_current (b[1]); + h2 = vlib_buffer_get_current (b[2]); + h3 = vlib_buffer_get_current (b[3]); - /* Save split horizon group */ - vnet_buffer (b0)->l2.shg = config->shg; + /* Save split horizon group */ + vnet_buffer (b[0])->l2.shg = config->shg; + vnet_buffer (b[1])->l2.shg = config->shg; + vnet_buffer (b[2])->l2.shg = config->shg; + vnet_buffer (b[3])->l2.shg = config->shg; - /* determine layer2 kind for stat and mask */ - if (PREDICT_FALSE (ethernet_address_cast (h0->dst_address))) - { - protocol = ((ip6_header_t *) l3h0)->protocol; + /* determine layer2 kind for stat and mask */ + if (PREDICT_FALSE (ethernet_address_cast (h0->dst_address))) + { + fm0 = l2_input_classiy_unicast (b[0], h0); + } + else + { + if (PREDICT_FALSE (sw_if_index == L2INPUT_BVI)) + vnet_buffer (b[0])->l2.shg = 0; + } + if (PREDICT_FALSE (ethernet_address_cast (h1->dst_address))) + { + fm1 = l2_input_classiy_unicast (b[1], h1); + } + else + { + if (PREDICT_FALSE (sw_if_index == L2INPUT_BVI)) + vnet_buffer (b[1])->l2.shg = 0; + } + if (PREDICT_FALSE (ethernet_address_cast (h2->dst_address))) + { + fm2 = l2_input_classiy_unicast (b[2], h2); + } + else + { + if (PREDICT_FALSE (sw_if_index == L2INPUT_BVI)) + vnet_buffer (b[1])->l2.shg = 0; + } + if (PREDICT_FALSE (ethernet_address_cast (h3->dst_address))) + { + fm3 = l2_input_classiy_unicast (b[3], h3); + } + else + { + if (PREDICT_FALSE (sw_if_index == L2INPUT_BVI)) + vnet_buffer (b[1])->l2.shg = 0; + } - /* Disable bridge forwarding (flooding will execute instead if not xconnect) */ - feat_mask &= ~(L2INPUT_FEAT_FWD | L2INPUT_FEAT_UU_FLOOD); + if (config->bridge) + { + /* Do bridge-domain processing */ + l2_bridge_domain_t *bd_config; + u16 bd_index; + + bd_index = config->bd_index; + bd_config = vec_elt_at_index (msm->bd_configs, bd_index); + + l2_input_classify_bridge (bd_config, bd_index, sw_if_index, + b[0], &fm0); + l2_input_classify_bridge (bd_config, bd_index, sw_if_index, + b[1], &fm1); + l2_input_classify_bridge (bd_config, bd_index, sw_if_index, + b[2], &fm2); + l2_input_classify_bridge (bd_config, bd_index, sw_if_index, + b[3], &fm3); + } + else if (config->xconnect) + { + /* Set the output interface */ + vnet_buffer (b[0])->sw_if_index[VLIB_TX] = + config->output_sw_if_index; + vnet_buffer (b[1])->sw_if_index[VLIB_TX] = + config->output_sw_if_index; + vnet_buffer (b[2])->sw_if_index[VLIB_TX] = + config->output_sw_if_index; + vnet_buffer (b[3])->sw_if_index[VLIB_TX] = + config->output_sw_if_index; + } + else + { + fm0 = L2INPUT_FEAT_DROP; + fm1 = L2INPUT_FEAT_DROP; + fm2 = L2INPUT_FEAT_DROP; + fm3 = L2INPUT_FEAT_DROP; + } - /* Disable ARP-term for non-ARP and non-ICMP6 packet */ - if (ethertype != ETHERNET_TYPE_ARP && - (ethertype != ETHERNET_TYPE_IP6 || protocol != IP_PROTOCOL_ICMP6)) - feat_mask &= ~(L2INPUT_FEAT_ARP_TERM); + /* mask out features from bitmap using packet type and bd config */ + fb0 = config->feature_bitmap & fm0; + fb1 = config->feature_bitmap & fm1; + fb2 = config->feature_bitmap & fm2; + fb3 = config->feature_bitmap & fm3; + + /* save for next feature graph nodes */ + vnet_buffer (b[0])->l2.feature_bitmap = fb0; + vnet_buffer (b[1])->l2.feature_bitmap = fb1; + vnet_buffer (b[2])->l2.feature_bitmap = fb2; + vnet_buffer (b[3])->l2.feature_bitmap = fb3; + + /* Determine the next node */ + *next = feat_bitmap_get_next_node_index (msm->feat_next_node_index, + fb0); + next++; + *next = feat_bitmap_get_next_node_index (msm->feat_next_node_index, + fb1); + next++; + *next = feat_bitmap_get_next_node_index (msm->feat_next_node_index, + fb2); + next++; + *next = feat_bitmap_get_next_node_index (msm->feat_next_node_index, + fb3); + next++; + + b += 4; + n_left -= 4; } - else + while (n_left) { - /* - * Check for from-BVI processing - set SHG of unicast packets from BVI - * to 0 so it is not dropped for VXLAN tunnels or other ports with the - * same SHG as that of the BVI. - */ - if (PREDICT_FALSE (vnet_buffer (b0)->sw_if_index[VLIB_TX] == - L2INPUT_BVI)) - vnet_buffer (b0)->l2.shg = 0; - } + const ethernet_header_t *h0; + u32 fm0, fb0; + fm0 = ~0; + h0 = vlib_buffer_get_current (b[0]); - if (config->xconnect) - { - /* Set the output interface */ - vnet_buffer (b0)->sw_if_index[VLIB_TX] = config->output_sw_if_index; - } - else - { - /* Do bridge-domain processing */ - bd_index0 = config->bd_index; - /* save BD ID for next feature graph nodes */ - vnet_buffer (b0)->l2.bd_index = bd_index0; + /* Save split horizon group */ + vnet_buffer (b[0])->l2.shg = config->shg; + + /* determine layer2 kind for stat and mask */ + if (PREDICT_FALSE (ethernet_address_cast (h0->dst_address))) + { + fm0 = l2_input_classiy_unicast (b[0], h0); + } + else + { + /* + * For packet from BVI - set SHG of unicast packet from BVI to 0 so it + * is not dropped on output to VXLAN tunnels or other ports with the + * same SHG as that of the BVI. + */ + if (PREDICT_FALSE (sw_if_index == L2INPUT_BVI)) + vnet_buffer (b[0])->l2.shg = 0; + } - /* Get config for the bridge domain interface */ - bd_config = vec_elt_at_index (msm->bd_configs, bd_index0); + if (config->bridge) + { + /* Do bridge-domain processing */ + u16 bd_index = config->bd_index; + l2_bridge_domain_t *bd_config = + vec_elt_at_index (msm->bd_configs, bd_index); - /* Save bridge domain and interface seq_num */ - /* *INDENT-OFF* */ - l2fib_seq_num_t sn = { - .swif = config->seq_num, - .bd = bd_config->seq_num, - }; - /* *INDENT-ON* */ - vnet_buffer (b0)->l2.l2fib_sn = sn.as_u16;; + l2_input_classify_bridge (bd_config, bd_index, sw_if_index, + b[0], &fm0); + } + else if (config->xconnect) + { + /* Set the output interface */ + vnet_buffer (b[0])->sw_if_index[VLIB_TX] = + config->output_sw_if_index; + } + else + fm0 = L2INPUT_FEAT_DROP; - /* - * Process bridge domain feature enables. - * To perform learning/flooding/forwarding, the corresponding bit - * must be enabled in both the input interface config and in the - * bridge domain config. In the bd_bitmap, bits for features other - * than learning/flooding/forwarding should always be set. - */ - feat_mask = feat_mask & bd_config->feature_bitmap; - } + /* mask out features from bitmap using packet type and bd config */ + fb0 = config->feature_bitmap & fm0; - /* mask out features from bitmap using packet type and bd config */ - feature_bitmap = config->feature_bitmap & feat_mask; + /* save for next feature graph nodes */ + vnet_buffer (b[0])->l2.feature_bitmap = fb0; - /* save for next feature graph nodes */ - vnet_buffer (b0)->l2.feature_bitmap = feature_bitmap; + /* Determine the next node */ + *next = feat_bitmap_get_next_node_index (msm->feat_next_node_index, + fb0); - /* Determine the next node */ - *next0 = feat_bitmap_get_next_node_index (msm->feat_next_node_index, - feature_bitmap); + next += 1; + b += 1; + n_left -= 1; + } } static_always_inline uword @@ -237,176 +435,91 @@ l2input_node_inline (vlib_main_t * vm, vlib_node_runtime_t * node, vlib_frame_t * frame, int do_trace) { - u32 n_left_from, *from, *to_next; - l2input_next_t next_index; + u32 n_left, *from; l2input_main_t *msm = &l2input_main; - u32 thread_index = vlib_get_thread_index (); + vlib_buffer_t *bufs[VLIB_FRAME_SIZE], **b; + u16 nexts[VLIB_FRAME_SIZE]; + u32 sw_if_indices[VLIB_FRAME_SIZE], *sw_if_index; from = vlib_frame_vector_args (frame); - n_left_from = frame->n_vectors; /* number of packets to process */ - next_index = node->cached_next_index; + n_left = frame->n_vectors; /* number of packets to process */ + vlib_get_buffers (vm, from, bufs, n_left); + b = bufs; + sw_if_index = sw_if_indices; - while (n_left_from > 0) + /* extract data from buffer metadata */ + while (n_left >= 8) + { + /* Prefetch the buffer header for the N+2 loop iteration */ + vlib_prefetch_buffer_header (b[4], LOAD); + vlib_prefetch_buffer_header (b[5], LOAD); + vlib_prefetch_buffer_header (b[6], LOAD); + vlib_prefetch_buffer_header (b[7], LOAD); + + sw_if_index[0] = vnet_buffer (b[0])->sw_if_index[VLIB_RX]; + sw_if_index[1] = vnet_buffer (b[1])->sw_if_index[VLIB_RX]; + sw_if_index[2] = vnet_buffer (b[2])->sw_if_index[VLIB_RX]; + sw_if_index[3] = vnet_buffer (b[3])->sw_if_index[VLIB_RX]; + + /* next */ + sw_if_index += 4; + n_left -= 4; + b += 4; + } + while (n_left) { - u32 n_left_to_next; + sw_if_index[0] = vnet_buffer (b[0])->sw_if_index[VLIB_RX]; - /* get space to enqueue frame to graph node "next_index" */ - vlib_get_next_frame (vm, node, next_index, to_next, n_left_to_next); + /* next */ + sw_if_index += 1; + n_left -= 1; + b += 1; + } - while (n_left_from >= 8 && n_left_to_next >= 4) - { - u32 bi0, bi1, bi2, bi3; - vlib_buffer_t *b0, *b1, *b2, *b3; - u32 next0, next1, next2, next3; - u32 sw_if_index0, sw_if_index1, sw_if_index2, sw_if_index3; - - /* Prefetch next iteration. */ - { - vlib_buffer_t *p4, *p5, *p6, *p7; - - p4 = vlib_get_buffer (vm, from[4]); - p5 = vlib_get_buffer (vm, from[5]); - p6 = vlib_get_buffer (vm, from[6]); - p7 = vlib_get_buffer (vm, from[7]); - - /* Prefetch the buffer header and packet for the N+2 loop iteration */ - vlib_prefetch_buffer_header (p4, LOAD); - vlib_prefetch_buffer_header (p5, LOAD); - vlib_prefetch_buffer_header (p6, LOAD); - vlib_prefetch_buffer_header (p7, LOAD); - - CLIB_PREFETCH (p4->data, CLIB_CACHE_LINE_BYTES, STORE); - CLIB_PREFETCH (p5->data, CLIB_CACHE_LINE_BYTES, STORE); - CLIB_PREFETCH (p6->data, CLIB_CACHE_LINE_BYTES, STORE); - CLIB_PREFETCH (p7->data, CLIB_CACHE_LINE_BYTES, STORE); - - /* - * Don't bother prefetching the bridge-domain config (which - * depends on the input config above). Only a small number of - * bridge domains are expected. Plus the structure is small - * and several fit in a cache line. - */ - } - - /* speculatively enqueue b0 and b1 to the current next frame */ - /* bi is "buffer index", b is pointer to the buffer */ - to_next[0] = bi0 = from[0]; - to_next[1] = bi1 = from[1]; - to_next[2] = bi2 = from[2]; - to_next[3] = bi3 = from[3]; - from += 4; - to_next += 4; - n_left_from -= 4; - n_left_to_next -= 4; - - b0 = vlib_get_buffer (vm, bi0); - b1 = vlib_get_buffer (vm, bi1); - b2 = vlib_get_buffer (vm, bi2); - b3 = vlib_get_buffer (vm, bi3); - - if (do_trace) - { - /* RX interface handles */ - sw_if_index0 = vnet_buffer (b0)->sw_if_index[VLIB_RX]; - sw_if_index1 = vnet_buffer (b1)->sw_if_index[VLIB_RX]; - sw_if_index2 = vnet_buffer (b2)->sw_if_index[VLIB_RX]; - sw_if_index3 = vnet_buffer (b3)->sw_if_index[VLIB_RX]; + n_left = frame->n_vectors; - if (b0->flags & VLIB_BUFFER_IS_TRACED) - { - ethernet_header_t *h0 = vlib_buffer_get_current (b0); - l2input_trace_t *t = - vlib_add_trace (vm, node, b0, sizeof (*t)); - t->sw_if_index = sw_if_index0; - clib_memcpy (t->src, h0->src_address, 6); - clib_memcpy (t->dst, h0->dst_address, 6); - } - if (b1->flags & VLIB_BUFFER_IS_TRACED) - { - ethernet_header_t *h1 = vlib_buffer_get_current (b1); - l2input_trace_t *t = - vlib_add_trace (vm, node, b1, sizeof (*t)); - t->sw_if_index = sw_if_index1; - clib_memcpy (t->src, h1->src_address, 6); - clib_memcpy (t->dst, h1->dst_address, 6); - } - if (b2->flags & VLIB_BUFFER_IS_TRACED) - { - ethernet_header_t *h2 = vlib_buffer_get_current (b2); - l2input_trace_t *t = - vlib_add_trace (vm, node, b2, sizeof (*t)); - t->sw_if_index = sw_if_index2; - clib_memcpy (t->src, h2->src_address, 6); - clib_memcpy (t->dst, h2->dst_address, 6); - } - if (b3->flags & VLIB_BUFFER_IS_TRACED) - { - ethernet_header_t *h3 = vlib_buffer_get_current (b3); - l2input_trace_t *t = - vlib_add_trace (vm, node, b3, sizeof (*t)); - t->sw_if_index = sw_if_index3; - clib_memcpy (t->src, h3->src_address, 6); - clib_memcpy (t->dst, h3->dst_address, 6); - } - } + while (n_left) + { + u16 count, *next; + u16 off = frame->n_vectors - n_left; + b = bufs + off; - vlib_node_increment_counter (vm, l2input_node.index, - L2INPUT_ERROR_L2INPUT, 4); + sw_if_index = sw_if_indices + off; + next = nexts + off; - classify_and_dispatch (vm, node, thread_index, msm, b0, &next0); - classify_and_dispatch (vm, node, thread_index, msm, b1, &next1); - classify_and_dispatch (vm, node, thread_index, msm, b2, &next2); - classify_and_dispatch (vm, node, thread_index, msm, b3, &next3); + count = clib_count_equal_u32 (sw_if_index, n_left); + n_left -= count; - /* verify speculative enqueues, maybe switch current next frame */ - /* if next0==next1==next_index then nothing special needs to be done */ - vlib_validate_buffer_enqueue_x4 (vm, node, next_index, - to_next, n_left_to_next, - bi0, bi1, bi2, bi3, - next0, next1, next2, next3); - } + classify_and_dispatch (msm, count, sw_if_index[0], b, next); + } - while (n_left_from > 0 && n_left_to_next > 0) + if (PREDICT_FALSE ((node->flags & VLIB_NODE_FLAG_TRACE))) + { + n_left = frame->n_vectors; /* number of packets to process */ + b = bufs; + + while (n_left) { - u32 bi0; - vlib_buffer_t *b0; - u32 next0; - u32 sw_if_index0; - - /* speculatively enqueue b0 to the current next frame */ - bi0 = from[0]; - to_next[0] = bi0; - from += 1; - to_next += 1; - n_left_from -= 1; - n_left_to_next -= 1; - - b0 = vlib_get_buffer (vm, bi0); - - if (do_trace && PREDICT_FALSE (b0->flags & VLIB_BUFFER_IS_TRACED)) + if (PREDICT_FALSE (b[0]->flags & VLIB_BUFFER_IS_TRACED)) { - ethernet_header_t *h0 = vlib_buffer_get_current (b0); - l2input_trace_t *t = vlib_add_trace (vm, node, b0, sizeof (*t)); - sw_if_index0 = vnet_buffer (b0)->sw_if_index[VLIB_RX]; - t->sw_if_index = sw_if_index0; + ethernet_header_t *h0 = vlib_buffer_get_current (b[0]); + l2input_trace_t *t = + vlib_add_trace (vm, node, b[0], sizeof (*t)); + + t->sw_if_index = vnet_buffer (b[0])->sw_if_index[VLIB_RX]; clib_memcpy (t->src, h0->src_address, 6); clib_memcpy (t->dst, h0->dst_address, 6); } - - vlib_node_increment_counter (vm, l2input_node.index, - L2INPUT_ERROR_L2INPUT, 1); - - classify_and_dispatch (vm, node, thread_index, msm, b0, &next0); - - /* verify speculative enqueue, maybe switch current next frame */ - vlib_validate_buffer_enqueue_x1 (vm, node, next_index, - to_next, n_left_to_next, - bi0, next0); + /* next */ + n_left--; + b++; } - - vlib_put_next_frame (vm, node, next_index, n_left_to_next); } + vlib_buffer_enqueue_to_next (vm, node, from, nexts, frame->n_vectors); + vlib_node_increment_counter (vm, l2input_node.index, + L2INPUT_ERROR_L2INPUT, frame->n_vectors); + return frame->n_vectors; } @@ -504,6 +617,27 @@ l2input_set_bridge_features (u32 bd_index, u32 feat_mask, u32 feat_value) return bd_config->feature_bitmap; } +void +l2input_interface_mac_change (u32 sw_if_index, + const u8 * old_address, const u8 * new_address) +{ + /* check if the sw_if_index passed is a BVI in a BD */ + l2_input_config_t *intf_config; + + intf_config = l2input_intf_config (sw_if_index); + + if (intf_config->bridge && intf_config->bvi) + { + /* delete and re-add l2fib entry for the bvi interface */ + l2fib_del_entry (old_address, intf_config->bd_index, sw_if_index); + l2fib_add_entry (new_address, + intf_config->bd_index, + sw_if_index, + L2FIB_ENTRY_RESULT_FLAG_BVI | + L2FIB_ENTRY_RESULT_FLAG_STATIC); + } +} + /** * Set the subinterface to run in l2 or l3 mode. * For L3 mode, just the sw_if_index is specified. @@ -528,9 +662,8 @@ set_int_l2_mode (vlib_main_t * vm, vnet_main_t * vnet_main, /* */ l2_output_config_t *out_config; l2_input_config_t *config; l2_bridge_domain_t *bd_config; - u64 mac; i32 l2_if_adjust = 0; - u32 slot; + vnet_device_class_t *dev_class; hi = vnet_get_sup_hw_interface (vnet_main, sw_if_index); config = l2input_intf_config (sw_if_index); @@ -546,19 +679,17 @@ set_int_l2_mode (vlib_main_t * vm, vnet_main_t * vnet_main, /* */ /* undo any BVI-related config */ if (bd_config->bvi_sw_if_index == sw_if_index) { + vnet_sw_interface_t *si; + bd_config->bvi_sw_if_index = ~0; config->bvi = 0; /* delete the l2fib entry for the bvi interface */ - mac = *((u64 *) hi->hw_address); - l2fib_del_entry (mac, config->bd_index); - - /* Make loop output node send packet back to ethernet-input node */ - slot = - vlib_node_add_named_next_with_slot (vm, hi->tx_node_index, - "ethernet-input", - VNET_SIMULATED_ETHERNET_TX_NEXT_ETHERNET_INPUT); - ASSERT (slot == VNET_SIMULATED_ETHERNET_TX_NEXT_ETHERNET_INPUT); + l2fib_del_entry (hi->hw_address, config->bd_index, sw_if_index); + + /* since this is a no longer BVI interface do not to flood to it */ + si = vnet_get_sw_interface (vnm, sw_if_index); + si->flood_class = VNET_FLOOD_CLASS_NO_FLOOD; } /* Clear MACs learned on the interface */ @@ -573,13 +704,9 @@ set_int_l2_mode (vlib_main_t * vm, vnet_main_t * vnet_main, /* */ l2_if_adjust--; } - /* - * Directs the l2 output path to work out the interface - * output next-arc itself. Needed when recycling a sw_if_index. - */ - vec_validate_init_empty (l2om->next_nodes.output_node_index_vec, - sw_if_index, ~0); - l2om->next_nodes.output_node_index_vec[sw_if_index] = ~0; + /* Make sure vector is big enough */ + vec_validate_init_empty (l2om->output_node_index_vec, sw_if_index, + L2OUTPUT_NEXT_DROP); /* Initialize the l2-input configuration for the interface */ if (mode == MODE_L3) @@ -598,29 +725,13 @@ set_int_l2_mode (vlib_main_t * vm, vnet_main_t * vnet_main, /* */ /* Make sure any L2-output packet to this interface now in L3 mode is * dropped. This may happen if L2 FIB MAC entry is stale */ - l2om->next_nodes.output_node_index_vec[sw_if_index] = - L2OUTPUT_NEXT_BAD_INTF; - } - else if (mode == MODE_L2_CLASSIFY) - { - config->xconnect = 1; - config->bridge = 0; - config->output_sw_if_index = xc_sw_if_index; - - /* Make sure last-chance drop is configured */ - config->feature_bitmap |= - L2INPUT_FEAT_DROP | L2INPUT_FEAT_INPUT_CLASSIFY; - - /* Make sure bridging features are disabled */ - config->feature_bitmap &= - ~(L2INPUT_FEAT_LEARN | L2INPUT_FEAT_FWD | L2INPUT_FEAT_FLOOD); - shg = 0; /* not used in xconnect */ - - /* Insure all packets go to ethernet-input */ - ethernet_set_rx_redirect (vnet_main, hi, 1); + l2om->output_node_index_vec[sw_if_index] = L2OUTPUT_NEXT_BAD_INTF; } else { + /* Add or update l2-output node next-arc and output_node_index_vec table + * for the interface */ + l2output_create_output_node_mapping (vm, vnet_main, sw_if_index); if (mode == MODE_L2_BRIDGE) { @@ -637,7 +748,7 @@ set_int_l2_mode (vlib_main_t * vm, vnet_main_t * vnet_main, /* */ config->xconnect = 0; config->bridge = 1; config->bd_index = bd_index; - config->seq_num += 1; + *l2fib_valid_swif_seq_num (sw_if_index) += 1; /* * Enable forwarding, flooding, learning and ARP termination by default @@ -661,6 +772,8 @@ set_int_l2_mode (vlib_main_t * vm, vnet_main_t * vnet_main, /* */ /* Do BVI interface initializations */ if (bvi) { + vnet_sw_interface_t *si; + /* ensure BD has no bvi interface (or replace that one with this??) */ if (bd_config->bvi_sw_if_index != ~0) { @@ -670,18 +783,16 @@ set_int_l2_mode (vlib_main_t * vm, vnet_main_t * vnet_main, /* */ config->bvi = 1; /* create the l2fib entry for the bvi interface */ - mac = *((u64 *) hi->hw_address); - l2fib_add_entry (mac, bd_index, sw_if_index, 1, 0, 1); /* static + bvi */ + l2fib_add_entry (hi->hw_address, bd_index, sw_if_index, + L2FIB_ENTRY_RESULT_FLAG_BVI | + L2FIB_ENTRY_RESULT_FLAG_STATIC); /* Disable learning by default. no use since l2fib entry is static. */ config->feature_bitmap &= ~L2INPUT_FEAT_LEARN; - /* Make loop output node send packet to l2-input node */ - slot = - vlib_node_add_named_next_with_slot (vm, hi->tx_node_index, - "l2-input", - VNET_SIMULATED_ETHERNET_TX_NEXT_ETHERNET_INPUT); - ASSERT (slot == VNET_SIMULATED_ETHERNET_TX_NEXT_ETHERNET_INPUT); + /* since this is a BVI interface we want to flood to it */ + si = vnet_get_sw_interface (vnm, sw_if_index); + si->flood_class = VNET_FLOOD_CLASS_BVI; } /* Add interface to bridge-domain flood vector */ @@ -693,7 +804,7 @@ set_int_l2_mode (vlib_main_t * vm, vnet_main_t * vnet_main, /* */ bd_add_member (bd_config, &member); } - else + else if (mode == MODE_L2_XC) { config->xconnect = 1; config->bridge = 0; @@ -709,6 +820,24 @@ set_int_l2_mode (vlib_main_t * vm, vnet_main_t * vnet_main, /* */ config->feature_bitmap |= L2INPUT_FEAT_XCONNECT; shg = 0; /* not used in xconnect */ } + else if (mode == MODE_L2_CLASSIFY) + { + config->xconnect = 1; + config->bridge = 0; + config->output_sw_if_index = xc_sw_if_index; + + /* Make sure last-chance drop is configured */ + config->feature_bitmap |= + L2INPUT_FEAT_DROP | L2INPUT_FEAT_INPUT_CLASSIFY; + + /* Make sure bridging features are disabled */ + config->feature_bitmap &= + ~(L2INPUT_FEAT_LEARN | L2INPUT_FEAT_FWD | L2INPUT_FEAT_FLOOD); + shg = 0; /* not used in xconnect */ + + /* Insure all packets go to ethernet-input */ + ethernet_set_rx_redirect (vnet_main, hi, 1); + } /* set up split-horizon group and set output feature bit */ config->shg = shg; @@ -757,6 +886,12 @@ set_int_l2_mode (vlib_main_t * vm, vnet_main_t * vnet_main, /* */ /* Set up the L2/L3 flag in the interface parsing tables */ ethernet_sw_interface_set_l2_mode (vnm, sw_if_index, (mode != MODE_L3)); + dev_class = vnet_get_device_class (vnet_main, hi->dev_class_index); + if (dev_class->set_l2_mode_function) + { + dev_class->set_l2_mode_function (vnet_main, hi, l2_if_adjust); + } + return 0; } @@ -1104,14 +1239,13 @@ _(l2fib_init) \ _(l2_input_classify_init) \ _(l2bd_init) \ _(l2fwd_init) \ -_(l2_inacl_init) \ +_(l2_in_out_acl_init) \ _(l2input_init) \ _(l2_vtr_init) \ _(l2_invtr_init) \ _(l2_efp_filter_init) \ _(l2learn_init) \ _(l2flood_init) \ -_(l2_outacl_init) \ _(l2output_init) \ _(l2_patch_init) \ _(l2_xcrw_init)