X-Git-Url: https://gerrit.fd.io/r/gitweb?a=blobdiff_plain;f=src%2Fvnet%2Fl2%2Fl2_output.c;h=31e3f067749b802147545ab322d271188358795e;hb=a1179589bea89b44f7b41001b5e5595a8793b1df;hp=acfe3abad904e575bc1eab62c393206bccf9669f;hpb=681abe46c9ea30e5c386db74c997714433470c4c;p=vpp.git diff --git a/src/vnet/l2/l2_output.c b/src/vnet/l2/l2_output.c index acfe3abad90..31e3f067749 100644 --- a/src/vnet/l2/l2_output.c +++ b/src/vnet/l2/l2_output.c @@ -27,6 +27,7 @@ #include +#ifndef CLIB_MARCH_VARIANT /* Feature graph node names */ static char *l2output_feat_names[] = { #define _(sym,name) name, @@ -40,7 +41,31 @@ l2output_get_feat_names (void) return l2output_feat_names; } +u8 * +format_l2_output_features (u8 * s, va_list * args) +{ + static char *display_names[] = { +#define _(sym,name) #sym, + foreach_l2output_feat +#undef _ + }; + u32 feature_bitmap = va_arg (*args, u32); + + if (feature_bitmap == 0) + { + s = format (s, " none configured"); + return s; + } + + int i; + for (i = L2OUTPUT_N_FEAT - 1; i >= 0; i--) + if (feature_bitmap & (1 << i)) + s = format (s, "%17s (%s)\n", display_names[i], l2output_feat_names[i]); + return s; +} + l2output_main_t l2output_main; +#endif typedef struct { @@ -48,6 +73,7 @@ typedef struct u8 src[6]; u8 dst[6]; u32 sw_if_index; + u8 raw[12]; /* raw data */ } l2output_trace_t; /* packet trace format function */ @@ -58,10 +84,15 @@ format_l2output_trace (u8 * s, va_list * args) CLIB_UNUSED (vlib_node_t * node) = va_arg (*args, vlib_node_t *); l2output_trace_t *t = va_arg (*args, l2output_trace_t *); - s = format (s, "l2-output: sw_if_index %d dst %U src %U", + s = format (s, "l2-output: sw_if_index %d dst %U src %U data " + "%02x %02x %02x %02x %02x %02x %02x %02x %02x %02x %02x %02x", t->sw_if_index, format_ethernet_address, t->dst, - format_ethernet_address, t->src); + format_ethernet_address, t->src, + t->raw[0], t->raw[1], t->raw[2], t->raw[3], t->raw[4], + t->raw[5], t->raw[6], t->raw[7], t->raw[8], t->raw[9], + t->raw[10], t->raw[11]); + return s; } @@ -79,372 +110,330 @@ static char *l2output_error_strings[] = { * split-horizon group as the input interface, except if the @c shg is 0 * in which case the check always passes. */ -static_always_inline u32 -split_horizon_violation (u8 shg1, u8 shg2) +static_always_inline void +split_horizon_violation (vlib_node_runtime_t * node, u8 shg, + vlib_buffer_t * b, u16 * next) { - if (PREDICT_TRUE (shg1 == 0)) - { - return 0; - } - else - { - return shg1 == shg2; - } + if (shg != vnet_buffer (b)->l2.shg) + return; + next[0] = L2OUTPUT_NEXT_DROP; + b->error = node->errors[L2OUTPUT_ERROR_SHG_DROP]; } static_always_inline void -l2output_vtr (vlib_node_runtime_t * node, l2_output_config_t * config, - u32 feature_bitmap, vlib_buffer_t * b, u32 * next) +l2output_process_batch_inline (vlib_main_t * vm, vlib_node_runtime_t * node, + l2_output_config_t * config, + vlib_buffer_t ** b, i16 * cdo, u16 * next, + u32 n_left, int l2_efp, int l2_vtr, int l2_pbb, + int shg_set, int update_feature_bitmap) { - if (PREDICT_FALSE (config->out_vtr_flag)) + while (n_left >= 8) { - /* Perform pre-vtr EFP filter check if configured */ - if (config->output_vtr.push_and_pop_bytes) + vlib_prefetch_buffer_header (b[4], LOAD); + vlib_prefetch_buffer_header (b[5], LOAD); + vlib_prefetch_buffer_header (b[6], LOAD); + vlib_prefetch_buffer_header (b[7], LOAD); + + /* prefetch eth headers only if we need to touch them */ + if (l2_vtr || l2_pbb || shg_set) { - /* - * Perform output vlan tag rewrite and the pre-vtr EFP filter check. - * The EFP Filter only needs to be run if there is an output VTR - * configured. The flag for the post-vtr EFP Filter node is used - * to trigger the pre-vtr check as well. - */ - u32 failed1 = (feature_bitmap & L2OUTPUT_FEAT_EFP_FILTER) - && (l2_efp_filter_process (b, &(config->input_vtr))); - u32 failed2 = l2_vtr_process (b, &(config->output_vtr)); + CLIB_PREFETCH (b[4]->data + cdo[4], CLIB_CACHE_LINE_BYTES, LOAD); + CLIB_PREFETCH (b[5]->data + cdo[5], CLIB_CACHE_LINE_BYTES, LOAD); + CLIB_PREFETCH (b[6]->data + cdo[6], CLIB_CACHE_LINE_BYTES, LOAD); + CLIB_PREFETCH (b[7]->data + cdo[7], CLIB_CACHE_LINE_BYTES, LOAD); + } - if (PREDICT_FALSE (failed1 | failed2)) + if (update_feature_bitmap) + { + vnet_buffer (b[0])->l2.feature_bitmap = config->feature_bitmap; + vnet_buffer (b[1])->l2.feature_bitmap = config->feature_bitmap; + vnet_buffer (b[2])->l2.feature_bitmap = config->feature_bitmap; + vnet_buffer (b[3])->l2.feature_bitmap = config->feature_bitmap; + } + + if (l2_vtr) + { + int i; + for (i = 0; i < 4; i++) { - *next = L2OUTPUT_NEXT_DROP; - if (failed2) + u32 failed1 = l2_efp && + l2_efp_filter_process (b[i], &(config->input_vtr)); + u32 failed2 = l2_vtr_process (b[i], &(config->output_vtr)); + if (PREDICT_FALSE (failed1 | failed2)) { - b->error = node->errors[L2OUTPUT_ERROR_VTR_DROP]; - } - if (failed1) - { - b->error = node->errors[L2OUTPUT_ERROR_EFP_DROP]; + next[i] = L2OUTPUT_NEXT_DROP; + if (failed2) + b[i]->error = node->errors[L2OUTPUT_ERROR_VTR_DROP]; + if (failed1) + b[i]->error = node->errors[L2OUTPUT_ERROR_EFP_DROP]; } } } - // perform the PBB rewrite - else if (config->output_pbb_vtr.push_and_pop_bytes) + + if (l2_pbb) + { + int i; + for (i = 0; i < 4; i++) + if (l2_pbb_process (b[i], &(config->output_pbb_vtr))) + { + next[i] = L2OUTPUT_NEXT_DROP; + b[i]->error = node->errors[L2OUTPUT_ERROR_VTR_DROP]; + } + } + + if (shg_set) + { + split_horizon_violation (node, config->shg, b[0], next); + split_horizon_violation (node, config->shg, b[1], next + 1); + split_horizon_violation (node, config->shg, b[2], next + 2); + split_horizon_violation (node, config->shg, b[3], next + 3); + } + /* next */ + n_left -= 4; + b += 4; + next += 4; + cdo += 4; + } + + while (n_left) + { + if (update_feature_bitmap) + vnet_buffer (b[0])->l2.feature_bitmap = config->feature_bitmap; + + if (l2_vtr) { - u32 failed = l2_pbb_process (b, &(config->output_pbb_vtr)); - if (PREDICT_FALSE (failed)) + u32 failed1 = l2_efp && + l2_efp_filter_process (b[0], &(config->input_vtr)); + u32 failed2 = l2_vtr_process (b[0], &(config->output_vtr)); + if (PREDICT_FALSE (failed1 | failed2)) { *next = L2OUTPUT_NEXT_DROP; - b->error = node->errors[L2OUTPUT_ERROR_VTR_DROP]; + if (failed2) + b[0]->error = node->errors[L2OUTPUT_ERROR_VTR_DROP]; + if (failed1) + b[0]->error = node->errors[L2OUTPUT_ERROR_EFP_DROP]; } } + + if (l2_pbb && l2_pbb_process (b[0], &(config->output_pbb_vtr))) + { + next[0] = L2OUTPUT_NEXT_DROP; + b[0]->error = node->errors[L2OUTPUT_ERROR_VTR_DROP]; + } + + if (shg_set) + split_horizon_violation (node, config->shg, b[0], next); + + /* next */ + n_left -= 1; + b += 1; + next += 1; } } +static_always_inline void +l2output_set_buffer_error (vlib_buffer_t ** b, u32 n_left, vlib_error_t error) +{ + while (n_left >= 8) + { + vlib_prefetch_buffer_header (b[4], LOAD); + vlib_prefetch_buffer_header (b[5], LOAD); + vlib_prefetch_buffer_header (b[6], LOAD); + vlib_prefetch_buffer_header (b[7], LOAD); + b[0]->error = b[1]->error = b[2]->error = b[3]->error = error; + b += 4; + n_left -= 4; + } + while (n_left) + { + b[0]->error = error; + b += 1; + n_left -= 1; + } +} -static vlib_node_registration_t l2output_node; +static_always_inline void +l2output_process_batch (vlib_main_t * vm, vlib_node_runtime_t * node, + l2_output_config_t * config, vlib_buffer_t ** b, + i16 * cdo, u16 * next, u32 n_left, int l2_efp, + int l2_vtr, int l2_pbb) +{ + u32 feature_bitmap = config->feature_bitmap & ~L2OUTPUT_FEAT_OUTPUT; + if (config->shg == 0 && feature_bitmap == 0) + { + if ((l2_efp | l2_vtr | l2_pbb) == 0) + return; + l2output_process_batch_inline (vm, node, config, b, cdo, next, n_left, + l2_efp, l2_vtr, l2_pbb, 0, 0); + } + else if (config->shg == 0) + l2output_process_batch_inline (vm, node, config, b, cdo, next, n_left, + l2_efp, l2_vtr, l2_pbb, 0, 1); + else if (feature_bitmap == 0) + l2output_process_batch_inline (vm, node, config, b, cdo, next, n_left, + l2_efp, l2_vtr, l2_pbb, 1, 0); + else + l2output_process_batch_inline (vm, node, config, b, cdo, next, n_left, + l2_efp, l2_vtr, l2_pbb, 1, 1); +} -static_always_inline uword -l2output_node_inline (vlib_main_t * vm, vlib_node_runtime_t * node, - vlib_frame_t * frame, int do_trace) +VLIB_NODE_FN (l2output_node) (vlib_main_t * vm, + vlib_node_runtime_t * node, + vlib_frame_t * frame) { - u32 n_left_from, *from, *to_next; - l2output_next_t next_index; + u32 n_left, *from; l2output_main_t *msm = &l2output_main; - u32 cached_sw_if_index; - u32 cached_next_index; - - /* Invalidate cache */ - cached_sw_if_index = ~0; - cached_next_index = ~0; /* warning be gone */ + vlib_buffer_t *bufs[VLIB_FRAME_SIZE], **b; + u16 nexts[VLIB_FRAME_SIZE]; + u32 sw_if_indices[VLIB_FRAME_SIZE], *sw_if_index; + i16 cur_data_offsets[VLIB_FRAME_SIZE], *cdo; + l2_output_config_t *config; + u32 feature_bitmap; from = vlib_frame_vector_args (frame); - n_left_from = frame->n_vectors; /* number of packets to process */ - next_index = node->cached_next_index; + n_left = frame->n_vectors; /* number of packets to process */ - while (n_left_from > 0) + vlib_get_buffers (vm, from, bufs, n_left); + b = bufs; + sw_if_index = sw_if_indices; + cdo = cur_data_offsets; + + /* extract data from buffer metadata */ + while (n_left >= 8) { - u32 n_left_to_next; + /* Prefetch the buffer header for the N+2 loop iteration */ + vlib_prefetch_buffer_header (b[4], LOAD); + vlib_prefetch_buffer_header (b[5], LOAD); + vlib_prefetch_buffer_header (b[6], LOAD); + vlib_prefetch_buffer_header (b[7], LOAD); + + sw_if_index[0] = vnet_buffer (b[0])->sw_if_index[VLIB_TX]; + cdo[0] = b[0]->current_data; + sw_if_index[1] = vnet_buffer (b[1])->sw_if_index[VLIB_TX]; + cdo[1] = b[1]->current_data; + sw_if_index[2] = vnet_buffer (b[2])->sw_if_index[VLIB_TX]; + cdo[2] = b[2]->current_data; + sw_if_index[3] = vnet_buffer (b[3])->sw_if_index[VLIB_TX]; + cdo[3] = b[3]->current_data; + + /* next */ + sw_if_index += 4; + n_left -= 4; + b += 4; + cdo += 4; + } + while (n_left) + { + sw_if_index[0] = vnet_buffer (b[0])->sw_if_index[VLIB_TX]; + cdo[0] = b[0]->current_data; + + /* next */ + sw_if_index += 1; + n_left -= 1; + b += 1; + cdo += 1; + } - /* get space to enqueue frame to graph node "next_index" */ - vlib_get_next_frame (vm, node, next_index, to_next, n_left_to_next); + n_left = frame->n_vectors; + while (n_left) + { + u16 count, new_next, *next; + u16 off = frame->n_vectors - n_left; + b = bufs + off; - while (n_left_from >= 8 && n_left_to_next >= 4) + if (n_left >= 4) { - u32 bi0, bi1, bi2, bi3; - vlib_buffer_t *b0, *b1, *b2, *b3; - u32 next0, next1, next2, next3; - u32 sw_if_index0, sw_if_index1, sw_if_index2, sw_if_index3; - ethernet_header_t *h0, *h1, *h2, *h3; - l2_output_config_t *config0, *config1, *config2, *config3; - u32 feature_bitmap0, feature_bitmap1; - u32 feature_bitmap2, feature_bitmap3; - - /* Prefetch next iteration. */ - { - vlib_buffer_t *p4, *p5, *p6, *p7; - - p4 = vlib_get_buffer (vm, from[4]); - p5 = vlib_get_buffer (vm, from[5]); - p6 = vlib_get_buffer (vm, from[6]); - p7 = vlib_get_buffer (vm, from[7]); - - /* Prefetch the buffer header for the N+2 loop iteration */ - vlib_prefetch_buffer_header (p4, LOAD); - vlib_prefetch_buffer_header (p5, LOAD); - vlib_prefetch_buffer_header (p6, LOAD); - vlib_prefetch_buffer_header (p7, LOAD); - } - - /* speculatively enqueue b0 and b1 to the current next frame */ - /* bi is "buffer index", b is pointer to the buffer */ - to_next[0] = bi0 = from[0]; - to_next[1] = bi1 = from[1]; - to_next[2] = bi2 = from[2]; - to_next[3] = bi3 = from[3]; - from += 4; - to_next += 4; - n_left_from -= 4; - n_left_to_next -= 4; - - b0 = vlib_get_buffer (vm, bi0); - b1 = vlib_get_buffer (vm, bi1); - b2 = vlib_get_buffer (vm, bi2); - b3 = vlib_get_buffer (vm, bi3); + vlib_prefetch_buffer_header (b[0], LOAD); + vlib_prefetch_buffer_header (b[1], LOAD); + vlib_prefetch_buffer_header (b[2], LOAD); + vlib_prefetch_buffer_header (b[3], LOAD); + } - /* TX interface handles */ - sw_if_index0 = vnet_buffer (b0)->sw_if_index[VLIB_TX]; - sw_if_index1 = vnet_buffer (b1)->sw_if_index[VLIB_TX]; - sw_if_index2 = vnet_buffer (b2)->sw_if_index[VLIB_TX]; - sw_if_index3 = vnet_buffer (b3)->sw_if_index[VLIB_TX]; + sw_if_index = sw_if_indices + off; + cdo = cur_data_offsets + off; + next = nexts + off; - if (do_trace) - { - h0 = vlib_buffer_get_current (b0); - h1 = vlib_buffer_get_current (b1); - h2 = vlib_buffer_get_current (b2); - h3 = vlib_buffer_get_current (b3); - if (b0->flags & VLIB_BUFFER_IS_TRACED) - { - l2output_trace_t *t = - vlib_add_trace (vm, node, b0, sizeof (*t)); - t->sw_if_index = sw_if_index0; - clib_memcpy (t->src, h0->src_address, 6); - clib_memcpy (t->dst, h0->dst_address, 6); - } - if (b1->flags & VLIB_BUFFER_IS_TRACED) - { - l2output_trace_t *t = - vlib_add_trace (vm, node, b1, sizeof (*t)); - t->sw_if_index = sw_if_index1; - clib_memcpy (t->src, h1->src_address, 6); - clib_memcpy (t->dst, h1->dst_address, 6); - } - if (b2->flags & VLIB_BUFFER_IS_TRACED) - { - l2output_trace_t *t = - vlib_add_trace (vm, node, b2, sizeof (*t)); - t->sw_if_index = sw_if_index2; - clib_memcpy (t->src, h2->src_address, 6); - clib_memcpy (t->dst, h2->dst_address, 6); - } - if (b3->flags & VLIB_BUFFER_IS_TRACED) - { - l2output_trace_t *t = - vlib_add_trace (vm, node, b3, sizeof (*t)); - t->sw_if_index = sw_if_index3; - clib_memcpy (t->src, h3->src_address, 6); - clib_memcpy (t->dst, h3->dst_address, 6); - } - } + count = clib_count_equal_u32 (sw_if_index, n_left); + n_left -= count; - vlib_node_increment_counter (vm, l2output_node.index, - L2OUTPUT_ERROR_L2OUTPUT, 4); - - /* Get config for the output interface */ - config0 = vec_elt_at_index (msm->configs, sw_if_index0); - config1 = vec_elt_at_index (msm->configs, sw_if_index1); - config2 = vec_elt_at_index (msm->configs, sw_if_index2); - config3 = vec_elt_at_index (msm->configs, sw_if_index3); - - /* - * Get features from the config - * TODO: mask out any non-applicable features - */ - feature_bitmap0 = config0->feature_bitmap; - feature_bitmap1 = config1->feature_bitmap; - feature_bitmap2 = config2->feature_bitmap; - feature_bitmap3 = config3->feature_bitmap; - - /* Determine next node */ - l2_output_dispatch (msm->vlib_main, - msm->vnet_main, - node, - l2output_node.index, - &cached_sw_if_index, - &cached_next_index, - &msm->next_nodes, - b0, sw_if_index0, feature_bitmap0, &next0); - - l2_output_dispatch (msm->vlib_main, - msm->vnet_main, - node, - l2output_node.index, - &cached_sw_if_index, - &cached_next_index, - &msm->next_nodes, - b1, sw_if_index1, feature_bitmap1, &next1); - - l2_output_dispatch (msm->vlib_main, - msm->vnet_main, - node, - l2output_node.index, - &cached_sw_if_index, - &cached_next_index, - &msm->next_nodes, - b2, sw_if_index2, feature_bitmap2, &next2); - - l2_output_dispatch (msm->vlib_main, - msm->vnet_main, - node, - l2output_node.index, - &cached_sw_if_index, - &cached_next_index, - &msm->next_nodes, - b3, sw_if_index3, feature_bitmap3, &next3); - - l2output_vtr (node, config0, feature_bitmap0, b0, &next0); - l2output_vtr (node, config1, feature_bitmap1, b1, &next1); - l2output_vtr (node, config2, feature_bitmap2, b2, &next2); - l2output_vtr (node, config3, feature_bitmap3, b3, &next3); - - /* - * Perform the split horizon check - * The check can only fail for non-zero shg's - */ - if (PREDICT_FALSE (config0->shg + config1->shg + - config2->shg + config3->shg)) - { - /* one of the checks might fail, check both */ - if (split_horizon_violation - (config0->shg, vnet_buffer (b0)->l2.shg)) - { - next0 = L2OUTPUT_NEXT_DROP; - b0->error = node->errors[L2OUTPUT_ERROR_SHG_DROP]; - } - if (split_horizon_violation - (config1->shg, vnet_buffer (b1)->l2.shg)) - { - next1 = L2OUTPUT_NEXT_DROP; - b1->error = node->errors[L2OUTPUT_ERROR_SHG_DROP]; - } - if (split_horizon_violation - (config2->shg, vnet_buffer (b2)->l2.shg)) - { - next2 = L2OUTPUT_NEXT_DROP; - b2->error = node->errors[L2OUTPUT_ERROR_SHG_DROP]; - } - if (split_horizon_violation - (config3->shg, vnet_buffer (b3)->l2.shg)) - { - next3 = L2OUTPUT_NEXT_DROP; - b3->error = node->errors[L2OUTPUT_ERROR_SHG_DROP]; - } - } + config = vec_elt_at_index (msm->configs, sw_if_index[0]); + feature_bitmap = config->feature_bitmap; + if (PREDICT_FALSE ((feature_bitmap & ~L2OUTPUT_FEAT_OUTPUT) != 0)) + new_next = feat_bitmap_get_next_node_index + (l2output_main.l2_out_feat_next, feature_bitmap); + else + new_next = vec_elt (l2output_main.output_node_index_vec, + sw_if_index[0]); + clib_memset_u16 (nexts + off, new_next, count); - /* verify speculative enqueues, maybe switch current next frame */ - /* if next0==next1==next_index then nothing special needs to be done */ - vlib_validate_buffer_enqueue_x4 (vm, node, next_index, - to_next, n_left_to_next, - bi0, bi1, bi2, bi3, - next0, next1, next2, next3); + if (new_next == L2OUTPUT_NEXT_DROP) + { + l2output_set_buffer_error + (b, count, node->errors[L2OUTPUT_ERROR_MAPPING_DROP]); + continue; } - while (n_left_from > 0 && n_left_to_next > 0) + /* VTR */ + if (config->out_vtr_flag && config->output_vtr.push_and_pop_bytes) { - u32 bi0; - vlib_buffer_t *b0; - u32 next0; - u32 sw_if_index0; - ethernet_header_t *h0; - l2_output_config_t *config0; - u32 feature_bitmap0; - - /* speculatively enqueue b0 to the current next frame */ - bi0 = from[0]; - to_next[0] = bi0; - from += 1; - to_next += 1; - n_left_from -= 1; - n_left_to_next -= 1; + if (feature_bitmap & L2OUTPUT_FEAT_EFP_FILTER) + l2output_process_batch (vm, node, config, b, cdo, next, count, + /* l2_efp */ 1, + /* l2_vtr */ 1, + /* l2_pbb */ 0); + else + l2output_process_batch (vm, node, config, b, cdo, next, count, + /* l2_efp */ 0, + /* l2_vtr */ 1, + /* l2_pbb */ 0); + } + else if (config->out_vtr_flag && + config->output_pbb_vtr.push_and_pop_bytes) + l2output_process_batch (vm, node, config, b, cdo, next, count, + /* l2_efp */ 0, + /* l2_vtr */ 0, + /* l2_pbb */ 1); + else + l2output_process_batch (vm, node, config, b, cdo, next, count, + /* l2_efp */ 0, + /* l2_vtr */ 0, + /* l2_pbb */ 0); + } - b0 = vlib_get_buffer (vm, bi0); - sw_if_index0 = vnet_buffer (b0)->sw_if_index[VLIB_TX]; + if (PREDICT_FALSE ((node->flags & VLIB_NODE_FLAG_TRACE))) + { + n_left = frame->n_vectors; /* number of packets to process */ + b = bufs; - if (do_trace && PREDICT_FALSE (b0->flags & VLIB_BUFFER_IS_TRACED)) + while (n_left) + { + if (PREDICT_FALSE (b[0]->flags & VLIB_BUFFER_IS_TRACED)) { + ethernet_header_t *h; l2output_trace_t *t = - vlib_add_trace (vm, node, b0, sizeof (*t)); - t->sw_if_index = sw_if_index0; - h0 = vlib_buffer_get_current (b0); - clib_memcpy (t->src, h0->src_address, 6); - clib_memcpy (t->dst, h0->dst_address, 6); + vlib_add_trace (vm, node, b[0], sizeof (*t)); + t->sw_if_index = vnet_buffer (b[0])->sw_if_index[VLIB_TX]; + h = vlib_buffer_get_current (b[0]); + clib_memcpy (t->src, h->src_address, 6); + clib_memcpy (t->dst, h->dst_address, 6); + clib_memcpy (t->raw, &h->type, sizeof (t->raw)); } - - vlib_node_increment_counter (vm, l2output_node.index, - L2OUTPUT_ERROR_L2OUTPUT, 1); - - /* Get config for the output interface */ - config0 = vec_elt_at_index (msm->configs, sw_if_index0); - - /* - * Get features from the config - * TODO: mask out any non-applicable features - */ - feature_bitmap0 = config0->feature_bitmap; - - /* Determine next node */ - l2_output_dispatch (msm->vlib_main, - msm->vnet_main, - node, - l2output_node.index, - &cached_sw_if_index, - &cached_next_index, - &msm->next_nodes, - b0, sw_if_index0, feature_bitmap0, &next0); - - l2output_vtr (node, config0, feature_bitmap0, b0, &next0); - - /* Perform the split horizon check */ - if (PREDICT_FALSE - (split_horizon_violation - (config0->shg, vnet_buffer (b0)->l2.shg))) - { - next0 = L2OUTPUT_NEXT_DROP; - b0->error = node->errors[L2OUTPUT_ERROR_SHG_DROP]; - } - - /* verify speculative enqueue, maybe switch current next frame */ - vlib_validate_buffer_enqueue_x1 (vm, node, next_index, - to_next, n_left_to_next, - bi0, next0); + /* next */ + n_left--; + b++; } - - vlib_put_next_frame (vm, node, next_index, n_left_to_next); } - return frame->n_vectors; -} + vlib_buffer_enqueue_to_next (vm, node, from, nexts, frame->n_vectors); + vlib_node_increment_counter (vm, l2output_node.index, + L2OUTPUT_ERROR_L2OUTPUT, frame->n_vectors); -static uword -l2output_node_fn (vlib_main_t * vm, - vlib_node_runtime_t * node, vlib_frame_t * frame) -{ - if (PREDICT_FALSE ((node->flags & VLIB_NODE_FLAG_TRACE))) - return l2output_node_inline (vm, node, frame, 1 /* do_trace */ ); - return l2output_node_inline (vm, node, frame, 0 /* do_trace */ ); + return frame->n_vectors; } /* *INDENT-OFF* */ -VLIB_REGISTER_NODE (l2output_node,static) = { - .function = l2output_node_fn, +VLIB_REGISTER_NODE (l2output_node) = { .name = "l2-output", .vector_size = sizeof (u32), .format_trace = format_l2output_trace, @@ -491,13 +480,12 @@ typedef enum * this sw_if_index, l2-output will send packets for this sw_if_index to the * l2-output-bad-intf node which just setup the proper drop reason before * sending packets to the error-drop node to drop the packet. Then, stale L2FIB - * entries for delted tunnels won't cause possible packet or memory corrpution. + * entries for deleted tunnels won't cause possible packet or memory corruption. */ -static vlib_node_registration_t l2output_bad_intf_node; -static uword -l2output_bad_intf_node_fn (vlib_main_t * vm, - vlib_node_runtime_t * node, vlib_frame_t * frame) +VLIB_NODE_FN (l2output_bad_intf_node) (vlib_main_t * vm, + vlib_node_runtime_t * node, + vlib_frame_t * frame) { u32 n_left_from, *from, *to_next; l2output_next_t next_index = 0; @@ -551,8 +539,7 @@ l2output_bad_intf_node_fn (vlib_main_t * vm, } /* *INDENT-OFF* */ -VLIB_REGISTER_NODE (l2output_bad_intf_node,static) = { - .function = l2output_bad_intf_node_fn, +VLIB_REGISTER_NODE (l2output_bad_intf_node) = { .name = "l2-output-bad-intf", .vector_size = sizeof (u32), .type = VLIB_NODE_TYPE_INTERNAL, @@ -569,9 +556,8 @@ VLIB_REGISTER_NODE (l2output_bad_intf_node,static) = { }; /* *INDENT-ON* */ - -VLIB_NODE_FUNCTION_MULTIARCH (l2output_node, l2output_node_fn) - clib_error_t *l2output_init (vlib_main_t * vm) +static clib_error_t * +l2output_init (vlib_main_t * vm) { l2output_main_t *mp = &l2output_main; @@ -587,93 +573,31 @@ VLIB_NODE_FUNCTION_MULTIARCH (l2output_node, l2output_node_fn) l2output_node.index, L2OUTPUT_N_FEAT, l2output_get_feat_names (), - mp->next_nodes.feat_next_node_index); + mp->l2_out_feat_next); /* Initialize the output node mapping table */ - l2output_init_output_node_vec (&mp->next_nodes.output_node_index_vec); + vec_validate_init_empty (mp->output_node_index_vec, 100, + L2OUTPUT_NEXT_DROP); return 0; } VLIB_INIT_FUNCTION (l2output_init); -typedef struct -{ - u32 node_index; - u32 sw_if_index; -} output_node_mapping_rpc_args_t; - -static void output_node_rpc_callback (output_node_mapping_rpc_args_t * a); - -static void -output_node_mapping_send_rpc (u32 node_index, u32 sw_if_index) -{ - output_node_mapping_rpc_args_t args; - void vl_api_rpc_call_main_thread (void *fp, u8 * data, u32 data_length); - - args.node_index = node_index; - args.sw_if_index = sw_if_index; - - vl_api_rpc_call_main_thread (output_node_rpc_callback, - (u8 *) & args, sizeof (args)); -} - +#ifndef CLIB_MARCH_VARIANT /** Create a mapping in the next node mapping table for the given sw_if_index. */ -u32 -l2output_create_output_node_mapping (vlib_main_t * vlib_main, vnet_main_t * vnet_main, u32 node_index, /* index of current node */ - u32 * output_node_index_vec, - u32 sw_if_index) -{ - - u32 next; /* index of next graph node */ - vnet_hw_interface_t *hw0; - u32 *node; - - hw0 = vnet_get_sup_hw_interface (vnet_main, sw_if_index); - - uword cpu_number; - - cpu_number = os_get_cpu_number (); - - if (cpu_number) - { - u32 oldflags; - - oldflags = __sync_fetch_and_or (&hw0->flags, - VNET_HW_INTERFACE_FLAG_L2OUTPUT_MAPPED); - - if ((oldflags & VNET_HW_INTERFACE_FLAG_L2OUTPUT_MAPPED)) - return L2OUTPUT_NEXT_DROP; - - output_node_mapping_send_rpc (node_index, sw_if_index); - return L2OUTPUT_NEXT_DROP; - } - - /* dynamically create graph node arc */ - next = vlib_node_add_next (vlib_main, node_index, hw0->output_node_index); - - /* Initialize vector with the mapping */ - - node = vec_elt_at_index (output_node_index_vec, sw_if_index); - *node = next; - - /* reset mapping bit, includes memory barrier */ - __sync_fetch_and_and (&hw0->flags, ~VNET_HW_INTERFACE_FLAG_L2OUTPUT_MAPPED); - - return next; -} - void -output_node_rpc_callback (output_node_mapping_rpc_args_t * a) +l2output_create_output_node_mapping (vlib_main_t * vlib_main, + vnet_main_t * vnet_main, u32 sw_if_index) { - vlib_main_t *vm = vlib_get_main (); - vnet_main_t *vnm = vnet_get_main (); - l2output_main_t *mp = &l2output_main; + vnet_hw_interface_t *hw0 = + vnet_get_sup_hw_interface (vnet_main, sw_if_index); - (void) l2output_create_output_node_mapping - (vm, vnm, a->node_index, mp->next_nodes.output_node_index_vec, - a->sw_if_index); + /* dynamically create graph node arc */ + u32 next = vlib_node_add_next (vlib_main, l2output_node.index, + hw0->output_node_index); + l2output_main.output_node_index_vec[sw_if_index] = next; } /* Get a pointer to the config for the given interface */ @@ -688,7 +612,8 @@ l2output_intf_config (u32 sw_if_index) /** Enable (or disable) the feature in the bitmap for the given interface. */ void -l2output_intf_bitmap_enable (u32 sw_if_index, u32 feature_bitmap, u32 enable) +l2output_intf_bitmap_enable (u32 sw_if_index, + l2output_feat_masks_t feature_bitmap, u32 enable) { l2output_main_t *mp = &l2output_main; l2_output_config_t *config; @@ -705,6 +630,7 @@ l2output_intf_bitmap_enable (u32 sw_if_index, u32 feature_bitmap, u32 enable) config->feature_bitmap &= ~feature_bitmap; } } +#endif /* * fd.io coding-style-patch-verification: ON