X-Git-Url: https://gerrit.fd.io/r/gitweb?a=blobdiff_plain;f=src%2Fvnet%2Fethernet%2Finterface.c;h=39e5cfb3cf89553e00efb081bec83baa45e8b32b;hb=cbe25aab3be72154f2c706c39eeba6a77f34450f;hp=a358080438a8f1ee6519a946f18b93ae65f4d403;hpb=635ec3bdd073101dd513083be4f1edfb64e29098;p=vpp.git diff --git a/src/vnet/ethernet/interface.c b/src/vnet/ethernet/interface.c index a358080438a..39e5cfb3cf8 100644 --- a/src/vnet/ethernet/interface.c +++ b/src/vnet/ethernet/interface.c @@ -41,8 +41,12 @@ #include #include #include +//#include #include +#include #include +#include +#include /** * @file @@ -51,7 +55,7 @@ * This file contains code to manage loopback interfaces. */ -const u8 * +static const u8 * ethernet_ip4_mcast_dst_addr (void) { const static u8 ethernet_mcast_dst_mac[] = { @@ -61,7 +65,7 @@ ethernet_ip4_mcast_dst_addr (void) return (ethernet_mcast_dst_mac); } -const u8 * +static const u8 * ethernet_ip6_mcast_dst_addr (void) { const static u8 ethernet_mcast_dst_mac[] = { @@ -151,7 +155,7 @@ ethernet_build_rewrite (vnet_main_t * vnm, if (dst_address) clib_memcpy (h->dst_address, dst_address, sizeof (h->dst_address)); else - memset (h->dst_address, ~0, sizeof (h->dst_address)); /* broadcast */ + clib_memset (h->dst_address, ~0, sizeof (h->dst_address)); /* broadcast */ } if (PREDICT_FALSE (!is_p2p) && sub_sw->sub.eth.flags.one_tag) @@ -193,32 +197,80 @@ ethernet_build_rewrite (vnet_main_t * vnm, void ethernet_update_adjacency (vnet_main_t * vnm, u32 sw_if_index, u32 ai) { - ip_adjacency_t *adj; - - adj = adj_get (ai); - vnet_sw_interface_t *si = vnet_get_sw_interface (vnm, sw_if_index); + if ((si->type == VNET_SW_INTERFACE_TYPE_P2P) || (si->type == VNET_SW_INTERFACE_TYPE_PIPE)) { default_update_adjacency (vnm, sw_if_index, ai); } - else if (FIB_PROTOCOL_IP4 == adj->ia_nh_proto) - { - arp_update_adjacency (vnm, sw_if_index, ai); - } - else if (FIB_PROTOCOL_IP6 == adj->ia_nh_proto) - { - ip6_ethernet_update_adjacency (vnm, sw_if_index, ai); - } else { - ASSERT (0); + ip_adjacency_t *adj; + + adj = adj_get (ai); + + switch (adj->lookup_next_index) + { + case IP_LOOKUP_NEXT_GLEAN: + adj_glean_update_rewrite (ai); + break; + case IP_LOOKUP_NEXT_ARP: + ip_neighbor_update (vnm, ai); + break; + case IP_LOOKUP_NEXT_BCAST: + adj_nbr_update_rewrite (ai, + ADJ_NBR_REWRITE_FLAG_COMPLETE, + ethernet_build_rewrite + (vnm, + adj->rewrite_header.sw_if_index, + adj->ia_link, + VNET_REWRITE_FOR_SW_INTERFACE_ADDRESS_BROADCAST)); + break; + case IP_LOOKUP_NEXT_MCAST: + { + /* + * Construct a partial rewrite from the known ethernet mcast dest MAC + */ + u8 *rewrite; + u8 offset; + + rewrite = ethernet_build_rewrite + (vnm, + sw_if_index, + adj->ia_link, + (adj->ia_nh_proto == FIB_PROTOCOL_IP6 ? + ethernet_ip6_mcast_dst_addr () : + ethernet_ip4_mcast_dst_addr ())); + + /* + * Complete the remaining fields of the adj's rewrite to direct the + * complete of the rewrite at switch time by copying in the IP + * dst address's bytes. + * Ofset is 2 bytes into the destintation address. + */ + offset = vec_len (rewrite) - 2; + adj_mcast_update_rewrite (ai, rewrite, offset); + + break; + } + case IP_LOOKUP_NEXT_DROP: + case IP_LOOKUP_NEXT_PUNT: + case IP_LOOKUP_NEXT_LOCAL: + case IP_LOOKUP_NEXT_REWRITE: + case IP_LOOKUP_NEXT_MCAST_MIDCHAIN: + case IP_LOOKUP_NEXT_MIDCHAIN: + case IP_LOOKUP_NEXT_ICMP_ERROR: + case IP_LOOKUP_N_NEXT: + ASSERT (0); + break; + } } } static clib_error_t * -ethernet_mac_change (vnet_hw_interface_t * hi, char *mac_address) +ethernet_mac_change (vnet_hw_interface_t * hi, + const u8 * old_address, const u8 * mac_address) { ethernet_interface_t *ei; ethernet_main_t *em; @@ -231,8 +283,12 @@ ethernet_mac_change (vnet_hw_interface_t * hi, char *mac_address) clib_memcpy (hi->hw_address, mac_address, vec_len (hi->hw_address)); clib_memcpy (ei->address, (u8 *) mac_address, sizeof (ei->address)); - ethernet_arp_change_mac (hi->sw_if_index); - ethernet_ndp_change_mac (hi->sw_if_index); + + { + ethernet_address_change_ctx_t *cb; + vec_foreach (cb, em->address_change_callbacks) + cb->function (em, hi->sw_if_index, cb->function_opaque); + } return (NULL); } @@ -275,7 +331,7 @@ clib_error_t * ethernet_register_interface (vnet_main_t * vnm, u32 dev_class_index, u32 dev_instance, - u8 * address, + const u8 * address, u32 * hw_if_index_return, ethernet_flag_change_function_t flag_change) { @@ -308,6 +364,7 @@ ethernet_register_interface (vnet_main_t * vnm, clib_memcpy (ei->address, address, sizeof (ei->address)); vec_add (hi->hw_address, address, sizeof (ei->address)); + CLIB_MEM_UNPOISON (hi->hw_address, 8); if (error) { @@ -340,6 +397,7 @@ ethernet_delete_interface (vnet_main_t * vnm, u32 hw_if_index) if (vlan_table->vlans[idx].qinqs) { pool_put_index (em->qinq_pool, vlan_table->vlans[idx].qinqs); + vlan_table->vlans[idx].qinqs = 0; } } pool_put_index (em->vlan_pool, main_intf->dot1q_vlans); @@ -353,6 +411,7 @@ ethernet_delete_interface (vnet_main_t * vnm, u32 hw_if_index) if (vlan_table->vlans[idx].qinqs) { pool_put_index (em->qinq_pool, vlan_table->vlans[idx].qinqs); + vlan_table->vlans[idx].qinqs = 0; } } pool_put_index (em->vlan_pool, main_intf->dot1ad_vlans); @@ -375,6 +434,7 @@ ethernet_set_flags (vnet_main_t * vnm, u32 hw_if_index, u32 flags) ASSERT (hi->hw_class_index == ethernet_hw_interface_class.index); ei = pool_elt_at_index (em->interfaces, hi->hw_instance); + ei->flags = flags; if (ei->flag_change) return ei->flag_change (vnm, hi, flags); return (u32) ~ 0; @@ -382,192 +442,236 @@ ethernet_set_flags (vnet_main_t * vnm, u32 hw_if_index, u32 flags) /** * Echo packets back to ethernet/l2-input. - * - * This node is "special." We know, for example, that - * all of the vnet_buffer (bX)->sw_if_index[VLIB_TX] values - * [had better!] match. - * - * Please do not copy the code first and ask questions later. - * - * "These are not the droids we're looking. - * You can go about your business. - * Move along..." */ static uword simulated_ethernet_interface_tx (vlib_main_t * vm, vlib_node_runtime_t * node, vlib_frame_t * frame) { - u32 n_left_from, n_left_to_next, *from, *to_next; - u32 next_index; - u32 n_bytes = 0; + u32 n_left_from, *from; + u32 next_index = 0; + u32 n_bytes; u32 thread_index = vm->thread_index; vnet_main_t *vnm = vnet_get_main (); vnet_interface_main_t *im = &vnm->interface_main; - u32 new_tx_sw_if_index, sw_if_index_all; - vlib_buffer_t *b0, *b1, *b2, *b3; - u32 bi0, bi1, bi2, bi3; - u32 next_all; l2_input_config_t *config; + vlib_buffer_t *bufs[VLIB_FRAME_SIZE], **b; + u16 nexts[VLIB_FRAME_SIZE], *next; + u32 new_rx_sw_if_index = ~0; + u32 new_tx_sw_if_index = ~0; n_left_from = frame->n_vectors; - from = vlib_frame_args (frame); + from = vlib_frame_vector_args (frame); - /* - * Work out where all of the packets are going. - */ - - bi0 = from[0]; - b0 = vlib_get_buffer (vm, bi0); - sw_if_index_all = vnet_buffer (b0)->sw_if_index[VLIB_TX]; + vlib_get_buffers (vm, from, bufs, n_left_from); + b = bufs; + next = nexts; - /* - * Look at the L2 config for the interface to decide which - * graph arc to use. If the interface is bridged, send pkts - * to l2-input. Otherwise, to ethernet-input - */ - config = l2input_intf_config (sw_if_index_all); - next_all = + /* Ordinarily, this is the only config lookup. */ + config = l2input_intf_config (vnet_buffer (b[0])->sw_if_index[VLIB_TX]); + next_index = config->bridge ? VNET_SIMULATED_ETHERNET_TX_NEXT_L2_INPUT : VNET_SIMULATED_ETHERNET_TX_NEXT_ETHERNET_INPUT; + new_tx_sw_if_index = config->bvi ? L2INPUT_BVI : ~0; + new_rx_sw_if_index = vnet_buffer (b[0])->sw_if_index[VLIB_TX]; - /* - * If the interface is a BVI, set the tx sw_if_index to the - * L2 path's special value. - * Otherwise, set it to ~0, to be reset later by the L3 path - */ - if (config->bvi) - new_tx_sw_if_index = L2INPUT_BVI; - else - new_tx_sw_if_index = ~0; - - /* Get the right next frame... */ - next_index = next_all; - - /* - * Use a quad-single loop, in case we have to impedance-match a - * full frame into a non-empty next frame or some such. - */ - - while (n_left_from > 0) + while (n_left_from >= 4) { - vlib_get_next_frame (vm, node, next_index, to_next, n_left_to_next); + u32 sw_if_index0, sw_if_index1, sw_if_index2, sw_if_index3; + u32 not_all_match_config; - while (n_left_from >= 4 && n_left_to_next >= 4) + /* Prefetch next iteration. */ + if (PREDICT_TRUE (n_left_from >= 8)) { - /* Prefetch next iteration. */ - if (PREDICT_TRUE (n_left_from >= 8)) - { - vlib_buffer_t *p4, *p5, *p6, *p7; - - p4 = vlib_get_buffer (vm, from[4]); - p5 = vlib_get_buffer (vm, from[5]); - p6 = vlib_get_buffer (vm, from[6]); - p7 = vlib_get_buffer (vm, from[7]); - - vlib_prefetch_buffer_header (p4, STORE); - vlib_prefetch_buffer_header (p5, STORE); - vlib_prefetch_buffer_header (p6, STORE); - vlib_prefetch_buffer_header (p7, STORE); - } - to_next[0] = bi0 = from[0]; - to_next[1] = bi1 = from[1]; - to_next[2] = bi2 = from[2]; - to_next[3] = bi3 = from[3]; - from += 4; - to_next += 4; - n_left_from -= 4; - n_left_to_next -= 4; - - b0 = vlib_get_buffer (vm, bi0); - b1 = vlib_get_buffer (vm, bi1); - b2 = vlib_get_buffer (vm, bi2); - b3 = vlib_get_buffer (vm, bi3); - - /* This "should never happen," of course... */ - if (CLIB_DEBUG > 0) - { - u32 cache_not_ok; - u32 sw_if_index0, sw_if_index1, sw_if_index2, sw_if_index3; - sw_if_index0 = vnet_buffer (b0)->sw_if_index[VLIB_TX]; - sw_if_index1 = vnet_buffer (b1)->sw_if_index[VLIB_TX]; - sw_if_index2 = vnet_buffer (b2)->sw_if_index[VLIB_TX]; - sw_if_index3 = vnet_buffer (b3)->sw_if_index[VLIB_TX]; - - cache_not_ok = (sw_if_index0 ^ sw_if_index1) - ^ (sw_if_index2 ^ sw_if_index3); - cache_not_ok += sw_if_index0 ^ sw_if_index_all; - ASSERT (cache_not_ok == 0); - } + vlib_prefetch_buffer_header (b[4], STORE); + vlib_prefetch_buffer_header (b[5], STORE); + vlib_prefetch_buffer_header (b[6], STORE); + vlib_prefetch_buffer_header (b[7], STORE); + } - vnet_buffer (b0)->sw_if_index[VLIB_RX] = sw_if_index_all; - vnet_buffer (b1)->sw_if_index[VLIB_RX] = sw_if_index_all; - vnet_buffer (b2)->sw_if_index[VLIB_RX] = sw_if_index_all; - vnet_buffer (b3)->sw_if_index[VLIB_RX] = sw_if_index_all; + /* Make sure all pkts were transmitted on the same (loop) intfc */ + sw_if_index0 = vnet_buffer (b[0])->sw_if_index[VLIB_TX]; + sw_if_index1 = vnet_buffer (b[1])->sw_if_index[VLIB_TX]; + sw_if_index2 = vnet_buffer (b[2])->sw_if_index[VLIB_TX]; + sw_if_index3 = vnet_buffer (b[3])->sw_if_index[VLIB_TX]; - vnet_buffer (b0)->sw_if_index[VLIB_TX] = new_tx_sw_if_index; - vnet_buffer (b1)->sw_if_index[VLIB_TX] = new_tx_sw_if_index; - vnet_buffer (b2)->sw_if_index[VLIB_TX] = new_tx_sw_if_index; - vnet_buffer (b3)->sw_if_index[VLIB_TX] = new_tx_sw_if_index; + not_all_match_config = (sw_if_index0 ^ sw_if_index1) + ^ (sw_if_index2 ^ sw_if_index3); + not_all_match_config += sw_if_index0 ^ new_rx_sw_if_index; - /* Update l2 lengths if necessary */ - if (next_all == VNET_SIMULATED_ETHERNET_TX_NEXT_L2_INPUT) + /* Speed path / expected case: all pkts on the same intfc */ + if (PREDICT_TRUE (not_all_match_config == 0)) + { + next[0] = next_index; + next[1] = next_index; + next[2] = next_index; + next[3] = next_index; + vnet_buffer (b[0])->sw_if_index[VLIB_RX] = new_rx_sw_if_index; + vnet_buffer (b[1])->sw_if_index[VLIB_RX] = new_rx_sw_if_index; + vnet_buffer (b[2])->sw_if_index[VLIB_RX] = new_rx_sw_if_index; + vnet_buffer (b[3])->sw_if_index[VLIB_RX] = new_rx_sw_if_index; + vnet_buffer (b[0])->sw_if_index[VLIB_TX] = new_tx_sw_if_index; + vnet_buffer (b[1])->sw_if_index[VLIB_TX] = new_tx_sw_if_index; + vnet_buffer (b[2])->sw_if_index[VLIB_TX] = new_tx_sw_if_index; + vnet_buffer (b[3])->sw_if_index[VLIB_TX] = new_tx_sw_if_index; + n_bytes = vlib_buffer_length_in_chain (vm, b[0]); + n_bytes += vlib_buffer_length_in_chain (vm, b[1]); + n_bytes += vlib_buffer_length_in_chain (vm, b[2]); + n_bytes += vlib_buffer_length_in_chain (vm, b[3]); + + if (next_index == VNET_SIMULATED_ETHERNET_TX_NEXT_L2_INPUT) { - vnet_update_l2_len (b0); - vnet_update_l2_len (b1); - vnet_update_l2_len (b2); - vnet_update_l2_len (b3); + vnet_update_l2_len (b[0]); + vnet_update_l2_len (b[1]); + vnet_update_l2_len (b[2]); + vnet_update_l2_len (b[3]); } - /* Byte accounting */ - n_bytes += vlib_buffer_length_in_chain (vm, b0); - n_bytes += vlib_buffer_length_in_chain (vm, b1); - n_bytes += vlib_buffer_length_in_chain (vm, b2); - n_bytes += vlib_buffer_length_in_chain (vm, b3); - - /* This *should* be a noop every time... */ - vlib_validate_buffer_enqueue_x4 (vm, node, next_index, - to_next, n_left_to_next, - bi0, bi1, bi2, bi3, - next_all, next_all, - next_all, next_all); + /* increment TX interface stat */ + vlib_increment_combined_counter (im->combined_sw_if_counters + + VNET_INTERFACE_COUNTER_TX, + thread_index, new_rx_sw_if_index, + 4 /* pkts */ , n_bytes); + b += 4; + next += 4; + n_left_from -= 4; + continue; } - while (n_left_from > 0 && n_left_to_next > 0) + /* + * Slow path: we know that at least one of the pkts + * was transmitted on a different sw_if_index, so + * check each sw_if_index against the cached data and proceed + * accordingly. + * + * This shouldn't happen, but code can (and does) bypass the + * per-interface output node, so deal with it. + */ + if (PREDICT_FALSE (vnet_buffer (b[0])->sw_if_index[VLIB_TX] + != new_rx_sw_if_index)) { - bi0 = from[0]; - to_next[0] = bi0; - from += 1; - to_next += 1; - n_left_from -= 1; - n_left_to_next -= 1; - - b0 = vlib_get_buffer (vm, bi0); - if (CLIB_DEBUG > 0) - { - u32 sw_if_index0 = vnet_buffer (b0)->sw_if_index[VLIB_TX]; - ASSERT (sw_if_index0 == sw_if_index_all); - } - - vnet_buffer (b0)->sw_if_index[VLIB_RX] = sw_if_index_all; - vnet_buffer (b0)->sw_if_index[VLIB_TX] = new_tx_sw_if_index; - if (next_all == VNET_SIMULATED_ETHERNET_TX_NEXT_L2_INPUT) - vnet_update_l2_len (b0); - - n_bytes += vlib_buffer_length_in_chain (vm, b0); - - vlib_validate_buffer_enqueue_x1 (vm, node, next_index, - to_next, n_left_to_next, - bi0, next_all); + config = l2input_intf_config + (vnet_buffer (b[0])->sw_if_index[VLIB_TX]); + next_index = + config->bridge ? VNET_SIMULATED_ETHERNET_TX_NEXT_L2_INPUT : + VNET_SIMULATED_ETHERNET_TX_NEXT_ETHERNET_INPUT; + new_tx_sw_if_index = config->bvi ? L2INPUT_BVI : ~0; + new_rx_sw_if_index = vnet_buffer (b[0])->sw_if_index[VLIB_TX]; + } + next[0] = next_index; + vnet_buffer (b[0])->sw_if_index[VLIB_RX] = new_rx_sw_if_index; + vnet_buffer (b[0])->sw_if_index[VLIB_TX] = new_tx_sw_if_index; + n_bytes = vlib_buffer_length_in_chain (vm, b[0]); + if (next_index == VNET_SIMULATED_ETHERNET_TX_NEXT_L2_INPUT) + vnet_update_l2_len (b[0]); + + vlib_increment_combined_counter (im->combined_sw_if_counters + + VNET_INTERFACE_COUNTER_TX, + thread_index, new_rx_sw_if_index, + 1 /* pkts */ , n_bytes); + + if (PREDICT_FALSE (vnet_buffer (b[1])->sw_if_index[VLIB_TX] + != new_rx_sw_if_index)) + { + config = l2input_intf_config + (vnet_buffer (b[1])->sw_if_index[VLIB_TX]); + next_index = + config->bridge ? VNET_SIMULATED_ETHERNET_TX_NEXT_L2_INPUT : + VNET_SIMULATED_ETHERNET_TX_NEXT_ETHERNET_INPUT; + new_rx_sw_if_index = vnet_buffer (b[1])->sw_if_index[VLIB_TX]; + new_tx_sw_if_index = config->bvi ? L2INPUT_BVI : ~0; + } + next[1] = next_index; + vnet_buffer (b[1])->sw_if_index[VLIB_RX] = new_rx_sw_if_index; + vnet_buffer (b[1])->sw_if_index[VLIB_TX] = new_tx_sw_if_index; + n_bytes = vlib_buffer_length_in_chain (vm, b[1]); + if (next_index == VNET_SIMULATED_ETHERNET_TX_NEXT_L2_INPUT) + vnet_update_l2_len (b[1]); + + vlib_increment_combined_counter (im->combined_sw_if_counters + + VNET_INTERFACE_COUNTER_TX, + thread_index, new_rx_sw_if_index, + 1 /* pkts */ , n_bytes); + + if (PREDICT_FALSE (vnet_buffer (b[2])->sw_if_index[VLIB_TX] + != new_rx_sw_if_index)) + { + config = l2input_intf_config + (vnet_buffer (b[2])->sw_if_index[VLIB_TX]); + next_index = + config->bridge ? VNET_SIMULATED_ETHERNET_TX_NEXT_L2_INPUT : + VNET_SIMULATED_ETHERNET_TX_NEXT_ETHERNET_INPUT; + new_rx_sw_if_index = vnet_buffer (b[2])->sw_if_index[VLIB_TX]; + new_tx_sw_if_index = config->bvi ? L2INPUT_BVI : ~0; } - vlib_put_next_frame (vm, node, next_index, n_left_to_next); + next[2] = next_index; + vnet_buffer (b[2])->sw_if_index[VLIB_RX] = new_rx_sw_if_index; + vnet_buffer (b[2])->sw_if_index[VLIB_TX] = new_tx_sw_if_index; + n_bytes = vlib_buffer_length_in_chain (vm, b[2]); + if (next_index == VNET_SIMULATED_ETHERNET_TX_NEXT_L2_INPUT) + vnet_update_l2_len (b[2]); + + vlib_increment_combined_counter (im->combined_sw_if_counters + + VNET_INTERFACE_COUNTER_TX, + thread_index, new_rx_sw_if_index, + 1 /* pkts */ , n_bytes); + + if (PREDICT_FALSE (vnet_buffer (b[3])->sw_if_index[VLIB_TX] + != new_rx_sw_if_index)) + { + config = l2input_intf_config + (vnet_buffer (b[3])->sw_if_index[VLIB_TX]); + next_index = + config->bridge ? VNET_SIMULATED_ETHERNET_TX_NEXT_L2_INPUT : + VNET_SIMULATED_ETHERNET_TX_NEXT_ETHERNET_INPUT; + new_rx_sw_if_index = vnet_buffer (b[3])->sw_if_index[VLIB_TX]; + new_tx_sw_if_index = config->bvi ? L2INPUT_BVI : ~0; + } + next[3] = next_index; + vnet_buffer (b[3])->sw_if_index[VLIB_RX] = new_rx_sw_if_index; + vnet_buffer (b[3])->sw_if_index[VLIB_TX] = new_tx_sw_if_index; + n_bytes = vlib_buffer_length_in_chain (vm, b[3]); + if (next_index == VNET_SIMULATED_ETHERNET_TX_NEXT_L2_INPUT) + vnet_update_l2_len (b[3]); + + vlib_increment_combined_counter (im->combined_sw_if_counters + + VNET_INTERFACE_COUNTER_TX, + thread_index, new_rx_sw_if_index, + 1 /* pkts */ , n_bytes); + b += 4; + next += 4; + n_left_from -= 4; + } + while (n_left_from > 0) + { + if (PREDICT_FALSE (vnet_buffer (b[0])->sw_if_index[VLIB_TX] + != new_rx_sw_if_index)) + { + config = l2input_intf_config + (vnet_buffer (b[0])->sw_if_index[VLIB_TX]); + next_index = + config->bridge ? VNET_SIMULATED_ETHERNET_TX_NEXT_L2_INPUT : + VNET_SIMULATED_ETHERNET_TX_NEXT_ETHERNET_INPUT; + new_tx_sw_if_index = config->bvi ? L2INPUT_BVI : ~0; + new_rx_sw_if_index = vnet_buffer (b[0])->sw_if_index[VLIB_TX]; + } + next[0] = next_index; + vnet_buffer (b[0])->sw_if_index[VLIB_RX] = new_rx_sw_if_index; + vnet_buffer (b[0])->sw_if_index[VLIB_TX] = new_tx_sw_if_index; + n_bytes = vlib_buffer_length_in_chain (vm, b[0]); + if (next_index == VNET_SIMULATED_ETHERNET_TX_NEXT_L2_INPUT) + vnet_update_l2_len (b[0]); + + vlib_increment_combined_counter (im->combined_sw_if_counters + + VNET_INTERFACE_COUNTER_TX, + thread_index, new_rx_sw_if_index, + 1 /* pkts */ , n_bytes); + b += 1; + next += 1; + n_left_from -= 1; } - /* increment TX interface stat */ - vlib_increment_combined_counter (im->combined_sw_if_counters + - VNET_INTERFACE_COUNTER_TX, - thread_index, sw_if_index_all, - frame->n_vectors, n_bytes); + vlib_buffer_enqueue_to_next (vm, node, from, nexts, frame->n_vectors); return frame->n_vectors; } @@ -589,18 +693,26 @@ simulated_ethernet_admin_up_down (vnet_main_t * vnm, u32 hw_if_index, return 0; } +static clib_error_t * +simulated_ethernet_mac_change (vnet_hw_interface_t * hi, + const u8 * old_address, const u8 * mac_address) +{ + l2input_interface_mac_change (hi->sw_if_index, old_address, mac_address); + + return (NULL); +} + + /* *INDENT-OFF* */ VNET_DEVICE_CLASS (ethernet_simulated_device_class) = { .name = "Loopback", .format_device_name = format_simulated_ethernet_name, .tx_function = simulated_ethernet_interface_tx, .admin_up_down_function = simulated_ethernet_admin_up_down, + .mac_addr_change_function = simulated_ethernet_mac_change, }; /* *INDENT-ON* */ -VLIB_DEVICE_TX_FUNCTION_MULTIARCH (ethernet_simulated_device_class, - simulated_ethernet_interface_tx); - /* * Maintain a bitmap of allocated loopback instance numbers. */ @@ -691,7 +803,7 @@ vnet_create_loopback_interface (u32 * sw_if_indexp, u8 * mac_address, *sw_if_indexp = (u32) ~ 0; - memset (address, 0, sizeof (address)); + clib_memset (address, 0, sizeof (address)); /* * Allocate a loopback instance. Either select on dynamically @@ -763,7 +875,7 @@ create_simulated_ethernet_interfaces (vlib_main_t * vm, u8 is_specified = 0; u32 user_instance = 0; - memset (mac_address, 0, sizeof (mac_address)); + clib_memset (mac_address, 0, sizeof (mac_address)); while (unformat_check_input (input) != UNFORMAT_END_OF_INPUT) { @@ -834,31 +946,113 @@ ethernet_get_interface (ethernet_main_t * em, u32 hw_if_index) index ? pool_elt_at_index (em->interfaces, i->hw_instance) : 0); } +mac_address_t * +ethernet_interface_add_del_address (ethernet_main_t * em, + u32 hw_if_index, const u8 * address, + u8 is_add) +{ + ethernet_interface_t *ei = ethernet_get_interface (em, hw_if_index); + mac_address_t *if_addr = 0; + + /* return if there is not an ethernet interface for this hw interface */ + if (!ei) + return 0; + + /* determine whether the address is configured on the interface */ + vec_foreach (if_addr, ei->secondary_addrs) + { + if (!ethernet_mac_address_equal (if_addr->bytes, address)) + continue; + + break; + } + + if (if_addr && vec_is_member (ei->secondary_addrs, if_addr)) + { + /* delete found address */ + if (!is_add) + { + vec_delete (ei->secondary_addrs, 1, if_addr - ei->secondary_addrs); + if_addr = 0; + } + /* address already found, so nothing needs to be done if adding */ + } + else + { + /* if_addr could be 0 or past the end of the vector. reset to 0 */ + if_addr = 0; + + /* add new address */ + if (is_add) + { + vec_add2 (ei->secondary_addrs, if_addr, 1); + clib_memcpy (&if_addr->bytes, address, sizeof (if_addr->bytes)); + } + } + + return if_addr; +} + int vnet_delete_loopback_interface (u32 sw_if_index) { vnet_main_t *vnm = vnet_get_main (); - vnet_sw_interface_t *si; - u32 hw_if_index; - vnet_hw_interface_t *hw; - u32 instance; if (pool_is_free_index (vnm->interface_main.sw_interfaces, sw_if_index)) return VNET_API_ERROR_INVALID_SW_IF_INDEX; - si = vnet_get_sw_interface (vnm, sw_if_index); - hw_if_index = si->hw_if_index; - hw = vnet_get_hw_interface (vnm, hw_if_index); - instance = hw->dev_instance; + vnet_hw_interface_t *hw = vnet_get_sup_hw_interface (vnm, sw_if_index); + if (hw == 0 || hw->dev_class_index != ethernet_simulated_device_class.index) + return VNET_API_ERROR_INVALID_SW_IF_INDEX; - if (loopback_instance_free (instance) < 0) + if (loopback_instance_free (hw->dev_instance) < 0) + return VNET_API_ERROR_INVALID_SW_IF_INDEX; + + ethernet_delete_interface (vnm, hw->hw_if_index); + + return 0; +} + +int +vnet_create_sub_interface (u32 sw_if_index, u32 id, + u32 flags, u16 inner_vlan_id, u16 outer_vlan_id, + u32 * sub_sw_if_index) +{ + vnet_main_t *vnm = vnet_get_main (); + vnet_interface_main_t *im = &vnm->interface_main; + vnet_hw_interface_t *hi; + u64 sup_and_sub_key = ((u64) (sw_if_index) << 32) | (u64) id; + vnet_sw_interface_t template; + uword *p; + u64 *kp; + + hi = vnet_get_sup_hw_interface (vnm, sw_if_index); + + p = hash_get_mem (im->sw_if_index_by_sup_and_sub, &sup_and_sub_key); + if (p) { - return VNET_API_ERROR_INVALID_SW_IF_INDEX; + return (VNET_API_ERROR_VLAN_ALREADY_EXISTS); } - ethernet_delete_interface (vnm, hw_if_index); + clib_memset (&template, 0, sizeof (template)); + template.type = VNET_SW_INTERFACE_TYPE_SUB; + template.flood_class = VNET_FLOOD_CLASS_NORMAL; + template.sup_sw_if_index = sw_if_index; + template.sub.id = id; + template.sub.eth.raw_flags = flags; + template.sub.eth.outer_vlan_id = outer_vlan_id; + template.sub.eth.inner_vlan_id = inner_vlan_id; - return 0; + if (vnet_create_sw_interface (vnm, &template, sub_sw_if_index)) + return (VNET_API_ERROR_UNSPECIFIED); + + kp = clib_mem_alloc (sizeof (*kp)); + *kp = sup_and_sub_key; + + hash_set (hi->sub_interface_sw_if_index_by_id, id, *sub_sw_if_index); + hash_set_mem (im->sw_if_index_by_sup_and_sub, kp, *sub_sw_if_index); + + return (0); } int