2 * Copyright (c) 2015 Cisco and/or its affiliates.
3 * Licensed under the Apache License, Version 2.0 (the "License");
4 * you may not use this file except in compliance with the License.
5 * You may obtain a copy of the License at:
7 * http://www.apache.org/licenses/LICENSE-2.0
9 * Unless required by applicable law or agreed to in writing, software
10 * distributed under the License is distributed on an "AS IS" BASIS,
11 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
12 * See the License for the specific language governing permissions and
13 * limitations under the License.
15 #include <vnet/l2/l2_xcrw.h>
19 * General L2 / L3 cross-connect, used to set up
20 * "L2 interface <--> your-favorite-tunnel-encap" tunnels.
22 * We set up a typical L2 cross-connect or (future) bridge
23 * to hook L2 interface(s) up to the L3 stack in arbitrary ways.
25 * Each l2_xcrw adjacency specifies 3 things:
27 * 1. The next graph node (presumably in the L3 stack) to
28 * process the (L2 -> L3) packet
30 * 2. A new value for vnet_buffer(b)->sw_if_index[VLIB_TX]
31 * (i.e. a lookup FIB index),
33 * 3. A rewrite string to apply.
35 * Example: to cross-connect an L2 interface or (future) bridge
36 * to an mpls-o-gre tunnel, set up the L2 rewrite string as shown in
37 * mpls_gre_rewrite, and use "mpls-post-rewrite" to fix the
38 * GRE IP header checksum and length fields.
47 /* packet trace format function */
49 format_l2_xcrw_trace (u8 * s, va_list * args)
51 CLIB_UNUSED (vlib_main_t * vm) = va_arg (*args, vlib_main_t *);
52 CLIB_UNUSED (vlib_node_t * node) = va_arg (*args, vlib_node_t *);
53 l2_xcrw_trace_t *t = va_arg (*args, l2_xcrw_trace_t *);
55 s = format (s, "L2_XCRW: next index %d tx_fib_index %d",
56 t->next_index, t->tx_fib_index);
60 extern l2_xcrw_main_t l2_xcrw_main;
62 #ifndef CLIB_MARCH_VARIANT
63 l2_xcrw_main_t l2_xcrw_main;
64 #endif /* CLIB_MARCH_VARIANT */
66 static char *l2_xcrw_error_strings[] = {
67 #define _(sym,string) string,
72 VLIB_NODE_FN (l2_xcrw_node) (vlib_main_t * vm,
73 vlib_node_runtime_t * node, vlib_frame_t * frame)
75 u32 n_left_from, *from, *to_next;
76 l2_xcrw_next_t next_index;
77 l2_xcrw_main_t *xcm = &l2_xcrw_main;
78 vlib_node_t *n = vlib_get_node (vm, l2_xcrw_node.index);
79 u32 node_counter_base_index = n->error_heap_index;
80 vlib_error_main_t *em = &vm->error_main;
82 from = vlib_frame_vector_args (frame);
83 n_left_from = frame->n_vectors;
84 next_index = node->cached_next_index;
86 while (n_left_from > 0)
90 vlib_get_next_frame (vm, node, next_index, to_next, n_left_to_next);
92 while (n_left_from >= 4 && n_left_to_next >= 2)
95 vlib_buffer_t *b0, *b1;
97 u32 sw_if_index0, sw_if_index1;
98 l2_xcrw_adjacency_t *adj0, *adj1;
100 /* Prefetch next iteration. */
102 vlib_buffer_t *p2, *p3;
104 p2 = vlib_get_buffer (vm, from[2]);
105 p3 = vlib_get_buffer (vm, from[3]);
107 vlib_prefetch_buffer_header (p2, LOAD);
108 vlib_prefetch_buffer_header (p3, LOAD);
110 CLIB_PREFETCH (p2->data, CLIB_CACHE_LINE_BYTES, STORE);
111 CLIB_PREFETCH (p3->data, CLIB_CACHE_LINE_BYTES, STORE);
114 /* speculatively enqueue b0 and b1 to the current next frame */
115 to_next[0] = bi0 = from[0];
116 to_next[1] = bi1 = from[1];
122 b0 = vlib_get_buffer (vm, bi0);
123 b1 = vlib_get_buffer (vm, bi1);
125 sw_if_index0 = vnet_buffer (b0)->sw_if_index[VLIB_RX];
126 sw_if_index1 = vnet_buffer (b1)->sw_if_index[VLIB_RX];
128 adj0 = vec_elt_at_index (xcm->adj_by_sw_if_index, sw_if_index0);
129 adj1 = vec_elt_at_index (xcm->adj_by_sw_if_index, sw_if_index1);
131 next0 = adj0->rewrite_header.next_index;
132 vnet_buffer (b0)->sw_if_index[VLIB_TX] =
133 adj0->rewrite_header.sw_if_index;
135 next1 = adj1->rewrite_header.next_index;
136 vnet_buffer (b1)->sw_if_index[VLIB_TX] =
137 adj1->rewrite_header.sw_if_index;
139 em->counters[node_counter_base_index + next1]++;
141 if (PREDICT_TRUE (next0 > 0))
143 u8 *h0 = vlib_buffer_get_current (b0);
144 vnet_rewrite_one_header (adj0[0], h0,
145 adj0->rewrite_header.data_bytes);
146 vlib_buffer_advance (b0, -adj0->rewrite_header.data_bytes);
147 em->counters[node_counter_base_index + L2_XCRW_ERROR_FWD]++;
150 if (PREDICT_TRUE (next1 > 0))
152 u8 *h1 = vlib_buffer_get_current (b1);
153 vnet_rewrite_one_header (adj1[0], h1,
154 adj1->rewrite_header.data_bytes);
155 vlib_buffer_advance (b1, -adj1->rewrite_header.data_bytes);
156 em->counters[node_counter_base_index + L2_XCRW_ERROR_FWD]++;
160 if (PREDICT_FALSE ((node->flags & VLIB_NODE_FLAG_TRACE)))
162 if (PREDICT_FALSE ((node->flags & VLIB_NODE_FLAG_TRACE)
163 && (b0->flags & VLIB_BUFFER_IS_TRACED)))
166 vlib_add_trace (vm, node, b0, sizeof (*t));
167 t->next_index = next0;
168 t->tx_fib_index = adj0->rewrite_header.sw_if_index;
170 if (PREDICT_FALSE ((node->flags & VLIB_NODE_FLAG_TRACE)
171 && (b1->flags & VLIB_BUFFER_IS_TRACED)))
174 vlib_add_trace (vm, node, b1, sizeof (*t));
175 t->next_index = next1;
176 t->tx_fib_index = adj1->rewrite_header.sw_if_index;
180 /* verify speculative enqueues, maybe switch current next frame */
181 vlib_validate_buffer_enqueue_x2 (vm, node, next_index,
182 to_next, n_left_to_next,
183 bi0, bi1, next0, next1);
186 while (n_left_from > 0 && n_left_to_next > 0)
192 l2_xcrw_adjacency_t *adj0;
194 /* speculatively enqueue b0 to the current next frame */
202 b0 = vlib_get_buffer (vm, bi0);
204 sw_if_index0 = vnet_buffer (b0)->sw_if_index[VLIB_RX];
206 adj0 = vec_elt_at_index (xcm->adj_by_sw_if_index, sw_if_index0);
208 next0 = adj0->rewrite_header.next_index;
209 vnet_buffer (b0)->sw_if_index[VLIB_TX] =
210 adj0->rewrite_header.sw_if_index;
212 if (PREDICT_TRUE (next0 > 0))
214 u8 *h0 = vlib_buffer_get_current (b0);
215 vnet_rewrite_one_header (adj0[0], h0,
216 adj0->rewrite_header.data_bytes);
217 vlib_buffer_advance (b0, -adj0->rewrite_header.data_bytes);
218 em->counters[node_counter_base_index + L2_XCRW_ERROR_FWD]++;
221 if (PREDICT_FALSE ((node->flags & VLIB_NODE_FLAG_TRACE)
222 && (b0->flags & VLIB_BUFFER_IS_TRACED)))
224 l2_xcrw_trace_t *t = vlib_add_trace (vm, node, b0, sizeof (*t));
225 t->next_index = next0;
226 t->tx_fib_index = adj0->rewrite_header.sw_if_index;
229 /* verify speculative enqueue, maybe switch current next frame */
230 vlib_validate_buffer_enqueue_x1 (vm, node, next_index,
231 to_next, n_left_to_next,
235 vlib_put_next_frame (vm, node, next_index, n_left_to_next);
238 return frame->n_vectors;
242 VLIB_REGISTER_NODE (l2_xcrw_node) = {
244 .vector_size = sizeof (u32),
245 .format_trace = format_l2_xcrw_trace,
246 .type = VLIB_NODE_TYPE_INTERNAL,
248 .n_errors = ARRAY_LEN(l2_xcrw_error_strings),
249 .error_strings = l2_xcrw_error_strings,
251 .n_next_nodes = L2_XCRW_N_NEXT,
253 /* edit / add dispositions here */
255 [L2_XCRW_NEXT_DROP] = "error-drop",
260 #ifndef CLIB_MARCH_VARIANT
262 l2_xcrw_init (vlib_main_t * vm)
264 l2_xcrw_main_t *mp = &l2_xcrw_main;
267 mp->vnet_main = &vnet_main;
268 mp->tunnel_index_by_l2_sw_if_index = hash_create (0, sizeof (uword));
273 VLIB_INIT_FUNCTION (l2_xcrw_init);
276 format_xcrw_name (u8 * s, va_list * args)
278 u32 dev_instance = va_arg (*args, u32);
279 return format (s, "xcrw%d", dev_instance);
283 VNET_DEVICE_CLASS (xcrw_device_class,static) = {
285 .format_device_name = format_xcrw_name,
289 /* Create a sham tunnel interface and return its sw_if_index */
291 create_xcrw_interface (vlib_main_t * vm)
293 vnet_main_t *vnm = vnet_get_main ();
297 vnet_hw_interface_t *hi;
300 /* mac address doesn't really matter */
301 clib_memset (address, 0, sizeof (address));
304 /* can returns error iff phy != 0 */
305 (void) ethernet_register_interface
306 (vnm, xcrw_device_class.index, instance++, address, &hw_if_index,
307 /* flag change */ 0);
309 hi = vnet_get_hw_interface (vnm, hw_if_index);
310 sw_if_index = hi->sw_if_index;
311 vnet_sw_interface_set_flags (vnm, sw_if_index,
312 VNET_SW_INTERFACE_FLAG_ADMIN_UP);
314 /* Output to the sham tunnel invokes the encap node */
315 hi->output_node_index = l2_xcrw_node.index;
321 vnet_configure_l2_xcrw (vlib_main_t * vm, vnet_main_t * vnm,
322 u32 l2_sw_if_index, u32 tx_fib_index,
323 u8 * rewrite, u32 next_node_index, int is_add)
325 l2_xcrw_main_t *xcm = &l2_xcrw_main;
326 l2_xcrw_adjacency_t *a;
333 pool_get (xcm->tunnels, t);
335 /* No interface allocated? Do it. Otherwise, set admin up */
336 if (t->tunnel_sw_if_index == 0)
337 t->tunnel_sw_if_index = create_xcrw_interface (vm);
339 vnet_sw_interface_set_flags (vnm, t->tunnel_sw_if_index,
340 VNET_SW_INTERFACE_FLAG_ADMIN_UP);
342 t->l2_sw_if_index = l2_sw_if_index;
344 vec_validate (xcm->adj_by_sw_if_index, t->l2_sw_if_index);
346 a = vec_elt_at_index (xcm->adj_by_sw_if_index, t->l2_sw_if_index);
347 clib_memset (a, 0, sizeof (*a));
349 a->rewrite_header.sw_if_index = tx_fib_index;
352 * Add or find a dynamic disposition for the successor node,
353 * e.g. so we can ship pkts to mpls_post_rewrite...
355 a->rewrite_header.next_index =
356 vlib_node_add_next (vm, l2_xcrw_node.index, next_node_index);
358 if (vec_len (rewrite))
359 vnet_rewrite_set_data (a[0], rewrite, vec_len (rewrite));
361 set_int_l2_mode (vm, vnm, MODE_L2_XC, t->l2_sw_if_index, 0,
362 L2_BD_PORT_TYPE_NORMAL, 0, t->tunnel_sw_if_index);
363 hash_set (xcm->tunnel_index_by_l2_sw_if_index,
364 t->l2_sw_if_index, t - xcm->tunnels);
369 p = hash_get (xcm->tunnel_index_by_l2_sw_if_index, l2_sw_if_index);
371 return VNET_API_ERROR_INVALID_SW_IF_INDEX;
373 t = pool_elt_at_index (xcm->tunnels, p[0]);
375 a = vec_elt_at_index (xcm->adj_by_sw_if_index, t->l2_sw_if_index);
376 /* Reset adj to drop traffic */
377 clib_memset (a, 0, sizeof (*a));
379 set_int_l2_mode (vm, vnm, MODE_L3, t->l2_sw_if_index, 0,
380 L2_BD_PORT_TYPE_NORMAL, 0, 0);
382 vnet_sw_interface_set_flags (vnm, t->tunnel_sw_if_index, 0 /* down */ );
384 hash_unset (xcm->tunnel_index_by_l2_sw_if_index, l2_sw_if_index);
385 pool_put (xcm->tunnels, t);
391 static clib_error_t *
392 set_l2_xcrw_command_fn (vlib_main_t * vm,
393 unformat_input_t * input, vlib_cli_command_t * cmd)
395 unformat_input_t _line_input, *line_input = &_line_input;
397 int is_ipv6 = 0; /* for fib id -> fib index mapping */
399 u32 tx_fib_index = ~0;
400 u32 next_node_index = ~0;
403 vnet_main_t *vnm = vnet_get_main ();
405 clib_error_t *error = NULL;
408 if (!unformat_user (input, unformat_line_input, line_input))
411 if (!unformat (line_input, "%U",
412 unformat_vnet_sw_interface, vnm, &l2_sw_if_index))
414 error = clib_error_return (0, "unknown input '%U'",
415 format_unformat_error, line_input);
419 while (unformat_check_input (line_input) != UNFORMAT_END_OF_INPUT)
421 if (unformat (line_input, "next %U",
422 unformat_vlib_node, vm, &next_node_index))
424 else if (unformat (line_input, "tx-fib-id %d", &tx_fib_id))
426 else if (unformat (line_input, "del"))
428 else if (unformat (line_input, "ipv6"))
430 else if (unformat (line_input, "rw %U", unformat_hex_string, &rw));
435 if (next_node_index == ~0)
437 error = clib_error_return (0, "next node not specified");
446 p = hash_get (ip6_main.fib_index_by_table_id, tx_fib_id);
448 p = hash_get (ip4_main.fib_index_by_table_id, tx_fib_id);
453 clib_error_return (0, "nonexistent tx_fib_id %d", tx_fib_id);
460 rv = vnet_configure_l2_xcrw (vm, vnm, l2_sw_if_index, tx_fib_index,
461 rw, next_node_index, is_add);
469 case VNET_API_ERROR_INVALID_SW_IF_INDEX:
470 error = clib_error_return (0, "%U not cross-connected",
471 format_vnet_sw_if_index_name,
472 vnm, l2_sw_if_index);
476 error = clib_error_return (0, "vnet_configure_l2_xcrw returned %d", rv);
482 unformat_free (line_input);
488 * Add or delete a Layer 2 to Layer 3 rewrite cross-connect. This is
489 * used to hook Layer 2 interface(s) up to the Layer 3 stack in
490 * arbitrary ways. For example, cross-connect an L2 interface or
491 * (future) bridge to an mpls-o-gre tunnel. Set up the L2 rewrite
492 * string as shown in mpls_gre_rewrite, and use \"mpls-post-rewrite\"
493 * to fix the GRE IP header checksum and length fields.
496 * @todo This is incomplete. This needs a detailed description and a
500 VLIB_CLI_COMMAND (set_l2_xcrw_command, static) = {
501 .path = "set interface l2 xcrw",
503 "set interface l2 xcrw <interface> next <node-name>\n"
504 " [del] [tx-fib-id <id>] [ipv6] rw <hex-bytes>",
505 .function = set_l2_xcrw_command_fn,
509 #endif /* CLIB_MARCH_VARIANT */
512 format_l2xcrw (u8 * s, va_list * args)
514 vnet_main_t *vnm = va_arg (*args, vnet_main_t *);
515 l2_xcrw_tunnel_t *t = va_arg (*args, l2_xcrw_tunnel_t *);
516 l2_xcrw_main_t *xcm = &l2_xcrw_main;
517 vlib_main_t *vm = vlib_get_main ();
518 l2_xcrw_adjacency_t *a;
523 s = format (s, "%-25s%s", "L2 interface", "Tunnel Details");
527 s = format (s, "%-25U %U ",
528 format_vnet_sw_if_index_name, vnm, t->l2_sw_if_index,
529 format_vnet_sw_if_index_name, vnm, t->tunnel_sw_if_index);
531 a = vec_elt_at_index (xcm->adj_by_sw_if_index, t->l2_sw_if_index);
533 s = format (s, "next %U ",
534 format_vlib_next_node_name, vm, l2_xcrw_node.index,
535 a->rewrite_header.next_index);
537 if (a->rewrite_header.sw_if_index != ~0)
538 s = format (s, "tx fib index %d ", a->rewrite_header.sw_if_index);
540 if (a->rewrite_header.data_bytes)
542 rewrite_string = (u8 *) (a + 1);
543 rewrite_string -= a->rewrite_header.data_bytes;
544 s = format (s, "rewrite data: %U ",
545 format_hex_bytes, rewrite_string,
546 a->rewrite_header.data_bytes);
549 s = format (s, "\n");
555 static clib_error_t *
556 show_l2xcrw_command_fn (vlib_main_t * vm,
557 unformat_input_t * input, vlib_cli_command_t * cmd)
559 vnet_main_t *vnm = vnet_get_main ();
560 l2_xcrw_main_t *xcm = &l2_xcrw_main;
563 if (pool_elts (xcm->tunnels) == 0)
565 vlib_cli_output (vm, "No L2 / L3 rewrite cross-connects configured");
569 vlib_cli_output (vm, "%U", format_l2xcrw, 0, 0);
572 pool_foreach (t, xcm->tunnels,
574 vlib_cli_output (vm, "%U", format_l2xcrw, vnm, t);
582 * Display a Layer 2 to Layer 3 rewrite cross-connect. This is used to
583 * hook Layer 2 interface(s) up to the Layer 3 stack in arbitrary ways.
585 * @todo This is incomplete. This needs a detailed description and a
589 VLIB_CLI_COMMAND (show_l2xcrw_command, static) = {
590 .path = "show l2xcrw",
591 .short_help = "show l2xcrw",
592 .function = show_l2xcrw_command_fn,
597 * fd.io coding-style-patch-verification: ON
600 * eval: (c-set-style "gnu")