dec0788c6530b1883232728c94243f248f7b5033
[vpp.git] / src / vnet / vxlan-gpe / decap.c
1 /*
2  * decap.c - decapsulate VXLAN GPE
3  *
4  * Copyright (c) 2013 Cisco and/or its affiliates.
5  * Licensed under the Apache License, Version 2.0 (the "License");
6  * you may not use this file except in compliance with the License.
7  * You may obtain a copy of the License at:
8  *
9  *     http://www.apache.org/licenses/LICENSE-2.0
10  *
11  * Unless required by applicable law or agreed to in writing, software
12  * distributed under the License is distributed on an "AS IS" BASIS,
13  * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
14  * See the License for the specific language governing permissions and
15  * limitations under the License.
16  */
17 /**
18  *  @file
19  *  @brief Functions for decapsulating VXLAN GPE tunnels
20  *
21 */
22
23 #include <vlib/vlib.h>
24 #include <vnet/pg/pg.h>
25 #include <vnet/vxlan-gpe/vxlan_gpe.h>
26
27 /**
28  * @brief Struct for VXLAN GPE decap packet tracing
29  *
30  */
31 typedef struct
32 {
33   u32 next_index;
34   u32 tunnel_index;
35   u32 error;
36 } vxlan_gpe_rx_trace_t;
37
38 /**
39  * @brief Tracing function for VXLAN GPE packet decapsulation
40  *
41  * @param *s
42  * @param *args
43  *
44  * @return *s
45  *
46  */
47 static u8 *
48 format_vxlan_gpe_rx_trace (u8 * s, va_list * args)
49 {
50   CLIB_UNUSED (vlib_main_t * vm) = va_arg (*args, vlib_main_t *);
51   CLIB_UNUSED (vlib_node_t * node) = va_arg (*args, vlib_node_t *);
52   vxlan_gpe_rx_trace_t *t = va_arg (*args, vxlan_gpe_rx_trace_t *);
53
54   if (t->tunnel_index != ~0)
55     {
56       s = format (s, "VXLAN-GPE: tunnel %d next %d error %d", t->tunnel_index,
57                   t->next_index, t->error);
58     }
59   else
60     {
61       s = format (s, "VXLAN-GPE: no tunnel next %d error %d\n", t->next_index,
62                   t->error);
63     }
64   return s;
65 }
66
67 /**
68  * @brief Tracing function for VXLAN GPE packet decapsulation including length
69  *
70  * @param *s
71  * @param *args
72  *
73  * @return *s
74  *
75  */
76 static u8 *
77 format_vxlan_gpe_with_length (u8 * s, va_list * args)
78 {
79   CLIB_UNUSED (vlib_main_t * vm) = va_arg (*args, vlib_main_t *);
80   CLIB_UNUSED (vlib_node_t * node) = va_arg (*args, vlib_node_t *);
81
82
83   return s;
84 }
85
86 /**
87  * @brief Common processing for IPv4 and IPv6 VXLAN GPE decap dispatch functions
88  *
89  * It is worth noting that other than trivial UDP forwarding (transit), VXLAN GPE
90  * tunnels are "terminate local". This means that there is no "TX" interface for this
91  * decap case, so that field in the buffer_metadata can be "used for something else".
92  * The something else in this case is, for the IPv4/IPv6 inner-packet type case, the
93  * FIB index used to look up the inner-packet's adjacency.
94  *
95  *      vnet_buffer(b0)->sw_if_index[VLIB_TX] = t0->decap_fib_index;
96  *
97  * @param *vm
98  * @param *node
99  * @param *from_frame
100  * @param is_ip4
101  *
102  * @return from_frame->n_vectors
103  *
104  */
105 always_inline uword
106 vxlan_gpe_input (vlib_main_t * vm,
107                  vlib_node_runtime_t * node,
108                  vlib_frame_t * from_frame, u8 is_ip4)
109 {
110   u32 n_left_from, next_index, *from, *to_next;
111   vxlan_gpe_main_t *nngm = &vxlan_gpe_main;
112   vnet_main_t *vnm = nngm->vnet_main;
113   vnet_interface_main_t *im = &vnm->interface_main;
114   u32 last_tunnel_index = ~0;
115   vxlan4_gpe_tunnel_key_t last_key4;
116   vxlan6_gpe_tunnel_key_t last_key6;
117   u32 pkts_decapsulated = 0;
118   u32 thread_index = vm->thread_index;
119   u32 stats_sw_if_index, stats_n_packets, stats_n_bytes;
120
121   if (is_ip4)
122     clib_memset (&last_key4, 0xff, sizeof (last_key4));
123   else
124     clib_memset (&last_key6, 0xff, sizeof (last_key6));
125
126   from = vlib_frame_vector_args (from_frame);
127   n_left_from = from_frame->n_vectors;
128
129   next_index = node->cached_next_index;
130   stats_sw_if_index = node->runtime_data[0];
131   stats_n_packets = stats_n_bytes = 0;
132
133   while (n_left_from > 0)
134     {
135       u32 n_left_to_next;
136
137       vlib_get_next_frame (vm, node, next_index, to_next, n_left_to_next);
138
139       while (n_left_from >= 4 && n_left_to_next >= 2)
140         {
141           u32 bi0, bi1;
142           vlib_buffer_t *b0, *b1;
143           u32 next0, next1;
144           ip4_vxlan_gpe_header_t *iuvn4_0, *iuvn4_1;
145           ip6_vxlan_gpe_header_t *iuvn6_0, *iuvn6_1;
146           uword *p0, *p1;
147           u32 tunnel_index0, tunnel_index1;
148           vxlan_gpe_tunnel_t *t0, *t1;
149           vxlan4_gpe_tunnel_key_t key4_0, key4_1;
150           vxlan6_gpe_tunnel_key_t key6_0, key6_1;
151           u32 error0, error1;
152           u32 sw_if_index0, sw_if_index1, len0, len1;
153
154           /* Prefetch next iteration. */
155           {
156             vlib_buffer_t *p2, *p3;
157
158             p2 = vlib_get_buffer (vm, from[2]);
159             p3 = vlib_get_buffer (vm, from[3]);
160
161             vlib_prefetch_buffer_header (p2, LOAD);
162             vlib_prefetch_buffer_header (p3, LOAD);
163
164             CLIB_PREFETCH (p2->data, 2 * CLIB_CACHE_LINE_BYTES, LOAD);
165             CLIB_PREFETCH (p3->data, 2 * CLIB_CACHE_LINE_BYTES, LOAD);
166           }
167
168           bi0 = from[0];
169           bi1 = from[1];
170           to_next[0] = bi0;
171           to_next[1] = bi1;
172           from += 2;
173           to_next += 2;
174           n_left_to_next -= 2;
175           n_left_from -= 2;
176
177           b0 = vlib_get_buffer (vm, bi0);
178           b1 = vlib_get_buffer (vm, bi1);
179
180           if (is_ip4)
181             {
182               /* udp leaves current_data pointing at the vxlan-gpe header */
183               vlib_buffer_advance (b0,
184                                    -(word) (sizeof (udp_header_t) +
185                                             sizeof (ip4_header_t)));
186               vlib_buffer_advance (b1,
187                                    -(word) (sizeof (udp_header_t) +
188                                             sizeof (ip4_header_t)));
189
190               iuvn4_0 = vlib_buffer_get_current (b0);
191               iuvn4_1 = vlib_buffer_get_current (b1);
192
193               /* pop (ip, udp, vxlan) */
194               vlib_buffer_advance (b0, sizeof (*iuvn4_0));
195               vlib_buffer_advance (b1, sizeof (*iuvn4_1));
196             }
197           else
198             {
199               /* udp leaves current_data pointing at the vxlan-gpe header */
200               vlib_buffer_advance (b0,
201                                    -(word) (sizeof (udp_header_t) +
202                                             sizeof (ip6_header_t)));
203               vlib_buffer_advance (b1,
204                                    -(word) (sizeof (udp_header_t) +
205                                             sizeof (ip6_header_t)));
206
207               iuvn6_0 = vlib_buffer_get_current (b0);
208               iuvn6_1 = vlib_buffer_get_current (b1);
209
210               /* pop (ip, udp, vxlan) */
211               vlib_buffer_advance (b0, sizeof (*iuvn6_0));
212               vlib_buffer_advance (b1, sizeof (*iuvn6_1));
213             }
214
215           tunnel_index0 = ~0;
216           tunnel_index1 = ~0;
217           error0 = 0;
218           error1 = 0;
219
220           if (is_ip4)
221             {
222               next0 =
223                 (iuvn4_0->vxlan.protocol < VXLAN_GPE_PROTOCOL_MAX) ?
224                 nngm->decap_next_node_list[iuvn4_0->vxlan.protocol] :
225                 VXLAN_GPE_INPUT_NEXT_DROP;
226               next1 =
227                 (iuvn4_1->vxlan.protocol < VXLAN_GPE_PROTOCOL_MAX) ?
228                 nngm->decap_next_node_list[iuvn4_1->vxlan.protocol] :
229                 VXLAN_GPE_INPUT_NEXT_DROP;
230
231               key4_0.local = iuvn4_0->ip4.dst_address.as_u32;
232               key4_1.local = iuvn4_1->ip4.dst_address.as_u32;
233
234               key4_0.remote = iuvn4_0->ip4.src_address.as_u32;
235               key4_1.remote = iuvn4_1->ip4.src_address.as_u32;
236
237               key4_0.vni = iuvn4_0->vxlan.vni_res;
238               key4_1.vni = iuvn4_1->vxlan.vni_res;
239
240               key4_0.pad = 0;
241               key4_1.pad = 0;
242             }
243           else                  /* is_ip6 */
244             {
245               next0 = (iuvn6_0->vxlan.protocol < node->n_next_nodes) ?
246                 iuvn6_0->vxlan.protocol : VXLAN_GPE_INPUT_NEXT_DROP;
247               next1 = (iuvn6_1->vxlan.protocol < node->n_next_nodes) ?
248                 iuvn6_1->vxlan.protocol : VXLAN_GPE_INPUT_NEXT_DROP;
249
250               key6_0.local.as_u64[0] = iuvn6_0->ip6.dst_address.as_u64[0];
251               key6_0.local.as_u64[1] = iuvn6_0->ip6.dst_address.as_u64[1];
252               key6_1.local.as_u64[0] = iuvn6_1->ip6.dst_address.as_u64[0];
253               key6_1.local.as_u64[1] = iuvn6_1->ip6.dst_address.as_u64[1];
254
255               key6_0.remote.as_u64[0] = iuvn6_0->ip6.src_address.as_u64[0];
256               key6_0.remote.as_u64[1] = iuvn6_0->ip6.src_address.as_u64[1];
257               key6_1.remote.as_u64[0] = iuvn6_1->ip6.src_address.as_u64[0];
258               key6_1.remote.as_u64[1] = iuvn6_1->ip6.src_address.as_u64[1];
259
260               key6_0.vni = iuvn6_0->vxlan.vni_res;
261               key6_1.vni = iuvn6_1->vxlan.vni_res;
262             }
263
264           /* Processing packet 0 */
265           if (is_ip4)
266             {
267               /* Processing for key4_0 */
268               if (PREDICT_FALSE ((key4_0.as_u64[0] != last_key4.as_u64[0])
269                                  || (key4_0.as_u64[1] !=
270                                      last_key4.as_u64[1])))
271                 {
272                   p0 = hash_get_mem (nngm->vxlan4_gpe_tunnel_by_key, &key4_0);
273
274                   if (p0 == 0)
275                     {
276                       error0 = VXLAN_GPE_ERROR_NO_SUCH_TUNNEL;
277                       goto trace0;
278                     }
279
280                   last_key4.as_u64[0] = key4_0.as_u64[0];
281                   last_key4.as_u64[1] = key4_0.as_u64[1];
282                   tunnel_index0 = last_tunnel_index = p0[0];
283                 }
284               else
285                 tunnel_index0 = last_tunnel_index;
286             }
287           else                  /* is_ip6 */
288             {
289               next0 =
290                 (iuvn6_0->vxlan.protocol < VXLAN_GPE_PROTOCOL_MAX) ?
291                 nngm->decap_next_node_list[iuvn6_0->vxlan.protocol] :
292                 VXLAN_GPE_INPUT_NEXT_DROP;
293               next1 =
294                 (iuvn6_1->vxlan.protocol < VXLAN_GPE_PROTOCOL_MAX) ?
295                 nngm->decap_next_node_list[iuvn6_1->vxlan.protocol] :
296                 VXLAN_GPE_INPUT_NEXT_DROP;
297
298               key6_0.local.as_u64[0] = iuvn6_0->ip6.dst_address.as_u64[0];
299               key6_0.local.as_u64[1] = iuvn6_0->ip6.dst_address.as_u64[1];
300               key6_1.local.as_u64[0] = iuvn6_1->ip6.dst_address.as_u64[0];
301               key6_1.local.as_u64[1] = iuvn6_1->ip6.dst_address.as_u64[1];
302
303               key6_0.remote.as_u64[0] = iuvn6_0->ip6.src_address.as_u64[0];
304               key6_0.remote.as_u64[1] = iuvn6_0->ip6.src_address.as_u64[1];
305               key6_1.remote.as_u64[0] = iuvn6_1->ip6.src_address.as_u64[0];
306               key6_1.remote.as_u64[1] = iuvn6_1->ip6.src_address.as_u64[1];
307
308               key6_0.vni = iuvn6_0->vxlan.vni_res;
309               key6_1.vni = iuvn6_1->vxlan.vni_res;
310
311               /* Processing for key6_0 */
312               if (PREDICT_FALSE
313                   (memcmp (&key6_0, &last_key6, sizeof (last_key6)) != 0))
314                 {
315                   p0 = hash_get_mem (nngm->vxlan6_gpe_tunnel_by_key, &key6_0);
316
317                   if (p0 == 0)
318                     {
319                       error0 = VXLAN_GPE_ERROR_NO_SUCH_TUNNEL;
320                       goto trace0;
321                     }
322
323                   memcpy (&last_key6, &key6_0, sizeof (key6_0));
324                   tunnel_index0 = last_tunnel_index = p0[0];
325                 }
326               else
327                 tunnel_index0 = last_tunnel_index;
328             }
329
330           t0 = pool_elt_at_index (nngm->tunnels, tunnel_index0);
331
332
333           sw_if_index0 = t0->sw_if_index;
334           len0 = vlib_buffer_length_in_chain (vm, b0);
335
336           /* Required to make the l2 tag push / pop code work on l2 subifs */
337           vnet_update_l2_len (b0);
338
339           /* Set packet input sw_if_index to unicast VXLAN tunnel for learning */
340           vnet_buffer (b0)->sw_if_index[VLIB_RX] = t0->sw_if_index;
341
342       /**
343        * ip[46] lookup in the configured FIB
344        */
345           vnet_buffer (b0)->sw_if_index[VLIB_TX] = t0->decap_fib_index;
346
347           pkts_decapsulated++;
348           stats_n_packets += 1;
349           stats_n_bytes += len0;
350
351           if (PREDICT_FALSE (sw_if_index0 != stats_sw_if_index))
352             {
353               stats_n_packets -= 1;
354               stats_n_bytes -= len0;
355               if (stats_n_packets)
356                 vlib_increment_combined_counter (im->combined_sw_if_counters +
357                                                  VNET_INTERFACE_COUNTER_RX,
358                                                  thread_index,
359                                                  stats_sw_if_index,
360                                                  stats_n_packets,
361                                                  stats_n_bytes);
362               stats_n_packets = 1;
363               stats_n_bytes = len0;
364               stats_sw_if_index = sw_if_index0;
365             }
366
367         trace0:b0->error = error0 ? node->errors[error0] : 0;
368
369           if (PREDICT_FALSE (b0->flags & VLIB_BUFFER_IS_TRACED))
370             {
371               vxlan_gpe_rx_trace_t *tr =
372                 vlib_add_trace (vm, node, b0, sizeof (*tr));
373               tr->next_index = next0;
374               tr->error = error0;
375               tr->tunnel_index = tunnel_index0;
376             }
377
378           /* Process packet 1 */
379           if (is_ip4)
380             {
381               /* Processing for key4_1 */
382               if (PREDICT_FALSE ((key4_1.as_u64[0] != last_key4.as_u64[0])
383                                  || (key4_1.as_u64[1] !=
384                                      last_key4.as_u64[1])))
385                 {
386                   p1 = hash_get_mem (nngm->vxlan4_gpe_tunnel_by_key, &key4_1);
387
388                   if (p1 == 0)
389                     {
390                       error1 = VXLAN_GPE_ERROR_NO_SUCH_TUNNEL;
391                       goto trace1;
392                     }
393
394                   last_key4.as_u64[0] = key4_1.as_u64[0];
395                   last_key4.as_u64[1] = key4_1.as_u64[1];
396                   tunnel_index1 = last_tunnel_index = p1[0];
397                 }
398               else
399                 tunnel_index1 = last_tunnel_index;
400             }
401           else                  /* is_ip6 */
402             {
403               /* Processing for key6_1 */
404               if (PREDICT_FALSE
405                   (memcmp (&key6_1, &last_key6, sizeof (last_key6)) != 0))
406                 {
407                   p1 = hash_get_mem (nngm->vxlan6_gpe_tunnel_by_key, &key6_1);
408
409                   if (p1 == 0)
410                     {
411                       error1 = VXLAN_GPE_ERROR_NO_SUCH_TUNNEL;
412                       goto trace1;
413                     }
414
415                   memcpy (&last_key6, &key6_1, sizeof (key6_1));
416                   tunnel_index1 = last_tunnel_index = p1[0];
417                 }
418               else
419                 tunnel_index1 = last_tunnel_index;
420             }
421
422           t1 = pool_elt_at_index (nngm->tunnels, tunnel_index1);
423
424           sw_if_index1 = t1->sw_if_index;
425           len1 = vlib_buffer_length_in_chain (vm, b1);
426
427           /* Required to make the l2 tag push / pop code work on l2 subifs */
428           vnet_update_l2_len (b1);
429
430           /* Set packet input sw_if_index to unicast VXLAN tunnel for learning */
431           vnet_buffer (b1)->sw_if_index[VLIB_RX] = t1->sw_if_index;
432
433           /*
434            * ip[46] lookup in the configured FIB
435            */
436           vnet_buffer (b1)->sw_if_index[VLIB_TX] = t1->decap_fib_index;
437
438           pkts_decapsulated++;
439           stats_n_packets += 1;
440           stats_n_bytes += len1;
441
442           /* Batch stats increment on the same vxlan tunnel so counter
443              is not incremented per packet */
444           if (PREDICT_FALSE (sw_if_index1 != stats_sw_if_index))
445             {
446               stats_n_packets -= 1;
447               stats_n_bytes -= len1;
448               if (stats_n_packets)
449                 vlib_increment_combined_counter (im->combined_sw_if_counters +
450                                                  VNET_INTERFACE_COUNTER_RX,
451                                                  thread_index,
452                                                  stats_sw_if_index,
453                                                  stats_n_packets,
454                                                  stats_n_bytes);
455               stats_n_packets = 1;
456               stats_n_bytes = len1;
457               stats_sw_if_index = sw_if_index1;
458             }
459           vnet_buffer (b1)->sw_if_index[VLIB_TX] = t1->decap_fib_index;
460
461         trace1:b1->error = error1 ? node->errors[error1] : 0;
462
463           if (PREDICT_FALSE (b1->flags & VLIB_BUFFER_IS_TRACED))
464             {
465               vxlan_gpe_rx_trace_t *tr =
466                 vlib_add_trace (vm, node, b1, sizeof (*tr));
467               tr->next_index = next1;
468               tr->error = error1;
469               tr->tunnel_index = tunnel_index1;
470             }
471
472           vlib_validate_buffer_enqueue_x2 (vm, node, next_index, to_next,
473                                            n_left_to_next, bi0, bi1, next0,
474                                            next1);
475         }
476
477       while (n_left_from > 0 && n_left_to_next > 0)
478         {
479           u32 bi0;
480           vlib_buffer_t *b0;
481           u32 next0;
482           ip4_vxlan_gpe_header_t *iuvn4_0;
483           ip6_vxlan_gpe_header_t *iuvn6_0;
484           uword *p0;
485           u32 tunnel_index0;
486           vxlan_gpe_tunnel_t *t0;
487           vxlan4_gpe_tunnel_key_t key4_0;
488           vxlan6_gpe_tunnel_key_t key6_0;
489           u32 error0;
490           u32 sw_if_index0, len0;
491
492           bi0 = from[0];
493           to_next[0] = bi0;
494           from += 1;
495           to_next += 1;
496           n_left_from -= 1;
497           n_left_to_next -= 1;
498
499           b0 = vlib_get_buffer (vm, bi0);
500
501           if (is_ip4)
502             {
503               /* udp leaves current_data pointing at the vxlan-gpe header */
504               vlib_buffer_advance (b0,
505                                    -(word) (sizeof (udp_header_t) +
506                                             sizeof (ip4_header_t)));
507
508               iuvn4_0 = vlib_buffer_get_current (b0);
509
510               /* pop (ip, udp, vxlan) */
511               vlib_buffer_advance (b0, sizeof (*iuvn4_0));
512             }
513           else
514             {
515               /* udp leaves current_data pointing at the vxlan-gpe header */
516               vlib_buffer_advance (b0,
517                                    -(word) (sizeof (udp_header_t) +
518                                             sizeof (ip6_header_t)));
519
520               iuvn6_0 = vlib_buffer_get_current (b0);
521
522               /* pop (ip, udp, vxlan) */
523               vlib_buffer_advance (b0, sizeof (*iuvn6_0));
524             }
525
526           tunnel_index0 = ~0;
527           error0 = 0;
528
529           if (is_ip4)
530             {
531               next0 =
532                 (iuvn4_0->vxlan.protocol < VXLAN_GPE_PROTOCOL_MAX) ?
533                 nngm->decap_next_node_list[iuvn4_0->vxlan.protocol] :
534                 VXLAN_GPE_INPUT_NEXT_DROP;
535
536               key4_0.local = iuvn4_0->ip4.dst_address.as_u32;
537               key4_0.remote = iuvn4_0->ip4.src_address.as_u32;
538               key4_0.vni = iuvn4_0->vxlan.vni_res;
539               key4_0.pad = 0;
540
541               /* Processing for key4_0 */
542               if (PREDICT_FALSE ((key4_0.as_u64[0] != last_key4.as_u64[0])
543                                  || (key4_0.as_u64[1] !=
544                                      last_key4.as_u64[1])))
545                 {
546                   p0 = hash_get_mem (nngm->vxlan4_gpe_tunnel_by_key, &key4_0);
547
548                   if (p0 == 0)
549                     {
550                       error0 = VXLAN_GPE_ERROR_NO_SUCH_TUNNEL;
551                       goto trace00;
552                     }
553
554                   last_key4.as_u64[0] = key4_0.as_u64[0];
555                   last_key4.as_u64[1] = key4_0.as_u64[1];
556                   tunnel_index0 = last_tunnel_index = p0[0];
557                 }
558               else
559                 tunnel_index0 = last_tunnel_index;
560             }
561           else                  /* is_ip6 */
562             {
563               next0 =
564                 (iuvn6_0->vxlan.protocol < VXLAN_GPE_PROTOCOL_MAX) ?
565                 nngm->decap_next_node_list[iuvn6_0->vxlan.protocol] :
566                 VXLAN_GPE_INPUT_NEXT_DROP;
567
568               key6_0.local.as_u64[0] = iuvn6_0->ip6.dst_address.as_u64[0];
569               key6_0.local.as_u64[1] = iuvn6_0->ip6.dst_address.as_u64[1];
570               key6_0.remote.as_u64[0] = iuvn6_0->ip6.src_address.as_u64[0];
571               key6_0.remote.as_u64[1] = iuvn6_0->ip6.src_address.as_u64[1];
572               key6_0.vni = iuvn6_0->vxlan.vni_res;
573
574               /* Processing for key6_0 */
575               if (PREDICT_FALSE
576                   (memcmp (&key6_0, &last_key6, sizeof (last_key6)) != 0))
577                 {
578                   p0 = hash_get_mem (nngm->vxlan6_gpe_tunnel_by_key, &key6_0);
579
580                   if (p0 == 0)
581                     {
582                       error0 = VXLAN_GPE_ERROR_NO_SUCH_TUNNEL;
583                       goto trace00;
584                     }
585
586                   memcpy (&last_key6, &key6_0, sizeof (key6_0));
587                   tunnel_index0 = last_tunnel_index = p0[0];
588                 }
589               else
590                 tunnel_index0 = last_tunnel_index;
591             }
592
593           t0 = pool_elt_at_index (nngm->tunnels, tunnel_index0);
594
595
596           sw_if_index0 = t0->sw_if_index;
597           len0 = vlib_buffer_length_in_chain (vm, b0);
598
599           /* Required to make the l2 tag push / pop code work on l2 subifs */
600           vnet_update_l2_len (b0);
601
602           /* Set packet input sw_if_index to unicast VXLAN tunnel for learning */
603           vnet_buffer (b0)->sw_if_index[VLIB_RX] = t0->sw_if_index;
604
605           /*
606            * ip[46] lookup in the configured FIB
607            */
608           vnet_buffer (b0)->sw_if_index[VLIB_TX] = t0->decap_fib_index;
609
610           pkts_decapsulated++;
611           stats_n_packets += 1;
612           stats_n_bytes += len0;
613
614           /* Batch stats increment on the same vxlan-gpe tunnel so counter
615              is not incremented per packet */
616           if (PREDICT_FALSE (sw_if_index0 != stats_sw_if_index))
617             {
618               stats_n_packets -= 1;
619               stats_n_bytes -= len0;
620               if (stats_n_packets)
621                 vlib_increment_combined_counter (im->combined_sw_if_counters +
622                                                  VNET_INTERFACE_COUNTER_RX,
623                                                  thread_index,
624                                                  stats_sw_if_index,
625                                                  stats_n_packets,
626                                                  stats_n_bytes);
627               stats_n_packets = 1;
628               stats_n_bytes = len0;
629               stats_sw_if_index = sw_if_index0;
630             }
631
632         trace00:b0->error = error0 ? node->errors[error0] : 0;
633
634           if (PREDICT_FALSE (b0->flags & VLIB_BUFFER_IS_TRACED))
635             {
636               vxlan_gpe_rx_trace_t *tr =
637                 vlib_add_trace (vm, node, b0, sizeof (*tr));
638               tr->next_index = next0;
639               tr->error = error0;
640               tr->tunnel_index = tunnel_index0;
641             }
642           vlib_validate_buffer_enqueue_x1 (vm, node, next_index, to_next,
643                                            n_left_to_next, bi0, next0);
644         }
645
646       vlib_put_next_frame (vm, node, next_index, n_left_to_next);
647     }
648
649   vlib_node_increment_counter (vm,
650                                is_ip4 ? vxlan4_gpe_input_node.index :
651                                vxlan6_gpe_input_node.index,
652                                VXLAN_GPE_ERROR_DECAPSULATED,
653                                pkts_decapsulated);
654
655   /* Increment any remaining batch stats */
656   if (stats_n_packets)
657     {
658       vlib_increment_combined_counter (im->combined_sw_if_counters +
659                                        VNET_INTERFACE_COUNTER_RX,
660                                        thread_index, stats_sw_if_index,
661                                        stats_n_packets, stats_n_bytes);
662       node->runtime_data[0] = stats_sw_if_index;
663     }
664   return from_frame->n_vectors;
665 }
666
667 /**
668  * @brief Graph processing dispatch function for IPv4 VXLAN GPE
669  *
670  * @node vxlan4-gpe-input
671  * @param *vm
672  * @param *node
673  * @param *from_frame
674  *
675  * @return from_frame->n_vectors
676  *
677  */
678 VLIB_NODE_FN (vxlan4_gpe_input_node) (vlib_main_t * vm,
679                                       vlib_node_runtime_t * node,
680                                       vlib_frame_t * from_frame)
681 {
682   return vxlan_gpe_input (vm, node, from_frame, /* is_ip4 */ 1);
683 }
684
685 #ifndef CLIB_MARCH_VARIANT
686 void
687 vxlan_gpe_register_decap_protocol (u8 protocol_id, uword next_node_index)
688 {
689   vxlan_gpe_main_t *hm = &vxlan_gpe_main;
690   hm->decap_next_node_list[protocol_id] = next_node_index;
691   return;
692 }
693
694 void
695 vxlan_gpe_unregister_decap_protocol (u8 protocol_id, uword next_node_index)
696 {
697   vxlan_gpe_main_t *hm = &vxlan_gpe_main;
698   hm->decap_next_node_list[protocol_id] = VXLAN_GPE_INPUT_NEXT_DROP;
699   return;
700 }
701 #endif /* CLIB_MARCH_VARIANT */
702
703 /**
704  * @brief Graph processing dispatch function for IPv6 VXLAN GPE
705  *
706  * @node vxlan6-gpe-input
707  * @param *vm
708  * @param *node
709  * @param *from_frame
710  *
711  * @return from_frame->n_vectors - uword
712  *
713  */
714 VLIB_NODE_FN (vxlan6_gpe_input_node) (vlib_main_t * vm,
715                                       vlib_node_runtime_t * node,
716                                       vlib_frame_t * from_frame)
717 {
718   return vxlan_gpe_input (vm, node, from_frame, /* is_ip4 */ 0);
719 }
720
721 /**
722  * @brief VXLAN GPE error strings
723  */
724 static char *vxlan_gpe_error_strings[] = {
725 #define vxlan_gpe_error(n,s) s,
726 #include <vnet/vxlan-gpe/vxlan_gpe_error.def>
727 #undef vxlan_gpe_error
728 #undef _
729 };
730
731 /* *INDENT-OFF* */
732 VLIB_REGISTER_NODE (vxlan4_gpe_input_node) = {
733   .name = "vxlan4-gpe-input",
734   /* Takes a vector of packets. */
735   .vector_size = sizeof (u32),
736   .type = VLIB_NODE_TYPE_INTERNAL,
737   .n_errors = ARRAY_LEN(vxlan_gpe_error_strings),
738   .error_strings = vxlan_gpe_error_strings,
739
740   .n_next_nodes = VXLAN_GPE_INPUT_N_NEXT,
741   .next_nodes = {
742 #define _(s,n) [VXLAN_GPE_INPUT_NEXT_##s] = n,
743     foreach_vxlan_gpe_input_next
744 #undef _
745   },
746
747   .format_buffer = format_vxlan_gpe_with_length,
748   .format_trace = format_vxlan_gpe_rx_trace,
749   // $$$$ .unformat_buffer = unformat_vxlan_gpe_header,
750 };
751 /* *INDENT-ON* */
752
753 /* *INDENT-OFF* */
754 VLIB_REGISTER_NODE (vxlan6_gpe_input_node) = {
755   .name = "vxlan6-gpe-input",
756   /* Takes a vector of packets. */
757   .vector_size = sizeof (u32),
758   .type = VLIB_NODE_TYPE_INTERNAL,
759   .n_errors = ARRAY_LEN(vxlan_gpe_error_strings),
760   .error_strings = vxlan_gpe_error_strings,
761
762   .n_next_nodes = VXLAN_GPE_INPUT_N_NEXT,
763   .next_nodes = {
764 #define _(s,n) [VXLAN_GPE_INPUT_NEXT_##s] = n,
765     foreach_vxlan_gpe_input_next
766 #undef _
767   },
768
769   .format_buffer = format_vxlan_gpe_with_length,
770   .format_trace = format_vxlan_gpe_rx_trace,
771   // $$$$ .unformat_buffer = unformat_vxlan_gpe_header,
772 };
773 /* *INDENT-ON* */
774
775 typedef enum
776 {
777   IP_VXLAN_BYPASS_NEXT_DROP,
778   IP_VXLAN_BYPASS_NEXT_VXLAN,
779   IP_VXLAN_BYPASS_N_NEXT,
780 } ip_vxlan_bypass_next_t;
781
782 always_inline uword
783 ip_vxlan_gpe_bypass_inline (vlib_main_t * vm,
784                             vlib_node_runtime_t * node,
785                             vlib_frame_t * frame, u32 is_ip4)
786 {
787   vxlan_gpe_main_t *ngm = &vxlan_gpe_main;
788   u32 *from, *to_next, n_left_from, n_left_to_next, next_index;
789   vlib_node_runtime_t *error_node =
790     vlib_node_get_runtime (vm, ip4_input_node.index);
791   ip4_address_t addr4;          /* last IPv4 address matching a local VTEP address */
792   ip6_address_t addr6;          /* last IPv6 address matching a local VTEP address */
793
794   from = vlib_frame_vector_args (frame);
795   n_left_from = frame->n_vectors;
796   next_index = node->cached_next_index;
797
798   if (node->flags & VLIB_NODE_FLAG_TRACE)
799     ip4_forward_next_trace (vm, node, frame, VLIB_TX);
800
801   if (is_ip4)
802     addr4.data_u32 = ~0;
803   else
804     ip6_address_set_zero (&addr6);
805
806   while (n_left_from > 0)
807     {
808       vlib_get_next_frame (vm, node, next_index, to_next, n_left_to_next);
809
810       while (n_left_from >= 4 && n_left_to_next >= 2)
811         {
812           vlib_buffer_t *b0, *b1;
813           ip4_header_t *ip40, *ip41;
814           ip6_header_t *ip60, *ip61;
815           udp_header_t *udp0, *udp1;
816           u32 bi0, ip_len0, udp_len0, flags0, next0;
817           u32 bi1, ip_len1, udp_len1, flags1, next1;
818           i32 len_diff0, len_diff1;
819           u8 error0, good_udp0, proto0;
820           u8 error1, good_udp1, proto1;
821
822           /* Prefetch next iteration. */
823           {
824             vlib_buffer_t *p2, *p3;
825
826             p2 = vlib_get_buffer (vm, from[2]);
827             p3 = vlib_get_buffer (vm, from[3]);
828
829             vlib_prefetch_buffer_header (p2, LOAD);
830             vlib_prefetch_buffer_header (p3, LOAD);
831
832             CLIB_PREFETCH (p2->data, 2 * CLIB_CACHE_LINE_BYTES, LOAD);
833             CLIB_PREFETCH (p3->data, 2 * CLIB_CACHE_LINE_BYTES, LOAD);
834           }
835
836           bi0 = to_next[0] = from[0];
837           bi1 = to_next[1] = from[1];
838           from += 2;
839           n_left_from -= 2;
840           to_next += 2;
841           n_left_to_next -= 2;
842
843           b0 = vlib_get_buffer (vm, bi0);
844           b1 = vlib_get_buffer (vm, bi1);
845           if (is_ip4)
846             {
847               ip40 = vlib_buffer_get_current (b0);
848               ip41 = vlib_buffer_get_current (b1);
849             }
850           else
851             {
852               ip60 = vlib_buffer_get_current (b0);
853               ip61 = vlib_buffer_get_current (b1);
854             }
855
856           /* Setup packet for next IP feature */
857           vnet_feature_next (&next0, b0);
858           vnet_feature_next (&next1, b1);
859
860           if (is_ip4)
861             {
862               proto0 = ip40->protocol;
863               proto1 = ip41->protocol;
864             }
865           else
866             {
867               proto0 = ip60->protocol;
868               proto1 = ip61->protocol;
869             }
870
871           /* Process packet 0 */
872           if (proto0 != IP_PROTOCOL_UDP)
873             goto exit0;         /* not UDP packet */
874
875           if (is_ip4)
876             udp0 = ip4_next_header (ip40);
877           else
878             udp0 = ip6_next_header (ip60);
879
880           if (udp0->dst_port != clib_host_to_net_u16 (UDP_DST_PORT_VXLAN_GPE))
881             goto exit0;         /* not VXLAN packet */
882
883           /* Validate DIP against VTEPs */
884           if (is_ip4)
885             {
886               if (addr4.as_u32 != ip40->dst_address.as_u32)
887                 {
888                   if (!hash_get (ngm->vtep4, ip40->dst_address.as_u32))
889                     goto exit0; /* no local VTEP for VXLAN packet */
890                   addr4 = ip40->dst_address;
891                 }
892             }
893           else
894             {
895               if (!ip6_address_is_equal (&addr6, &ip60->dst_address))
896                 {
897                   if (!hash_get_mem (ngm->vtep6, &ip60->dst_address))
898                     goto exit0; /* no local VTEP for VXLAN packet */
899                   addr6 = ip60->dst_address;
900                 }
901             }
902
903           flags0 = b0->flags;
904           good_udp0 = (flags0 & VNET_BUFFER_F_L4_CHECKSUM_CORRECT) != 0;
905
906           /* Don't verify UDP checksum for packets with explicit zero checksum. */
907           good_udp0 |= udp0->checksum == 0;
908
909           /* Verify UDP length */
910           if (is_ip4)
911             ip_len0 = clib_net_to_host_u16 (ip40->length);
912           else
913             ip_len0 = clib_net_to_host_u16 (ip60->payload_length);
914           udp_len0 = clib_net_to_host_u16 (udp0->length);
915           len_diff0 = ip_len0 - udp_len0;
916
917           /* Verify UDP checksum */
918           if (PREDICT_FALSE (!good_udp0))
919             {
920               if ((flags0 & VNET_BUFFER_F_L4_CHECKSUM_COMPUTED) == 0)
921                 {
922                   if (is_ip4)
923                     flags0 = ip4_tcp_udp_validate_checksum (vm, b0);
924                   else
925                     flags0 = ip6_tcp_udp_icmp_validate_checksum (vm, b0);
926                   good_udp0 =
927                     (flags0 & VNET_BUFFER_F_L4_CHECKSUM_CORRECT) != 0;
928                 }
929             }
930
931           if (is_ip4)
932             {
933               error0 = good_udp0 ? 0 : IP4_ERROR_UDP_CHECKSUM;
934               error0 = (len_diff0 >= 0) ? error0 : IP4_ERROR_UDP_LENGTH;
935             }
936           else
937             {
938               error0 = good_udp0 ? 0 : IP6_ERROR_UDP_CHECKSUM;
939               error0 = (len_diff0 >= 0) ? error0 : IP6_ERROR_UDP_LENGTH;
940             }
941
942           next0 = error0 ?
943             IP_VXLAN_BYPASS_NEXT_DROP : IP_VXLAN_BYPASS_NEXT_VXLAN;
944           b0->error = error0 ? error_node->errors[error0] : 0;
945
946           /* vxlan_gpe-input node expect current at VXLAN header */
947           if (is_ip4)
948             vlib_buffer_advance (b0,
949                                  sizeof (ip4_header_t) +
950                                  sizeof (udp_header_t));
951           else
952             vlib_buffer_advance (b0,
953                                  sizeof (ip6_header_t) +
954                                  sizeof (udp_header_t));
955
956         exit0:
957           /* Process packet 1 */
958           if (proto1 != IP_PROTOCOL_UDP)
959             goto exit1;         /* not UDP packet */
960
961           if (is_ip4)
962             udp1 = ip4_next_header (ip41);
963           else
964             udp1 = ip6_next_header (ip61);
965
966           if (udp1->dst_port != clib_host_to_net_u16 (UDP_DST_PORT_VXLAN_GPE))
967             goto exit1;         /* not VXLAN packet */
968
969           /* Validate DIP against VTEPs */
970           if (is_ip4)
971             {
972               if (addr4.as_u32 != ip41->dst_address.as_u32)
973                 {
974                   if (!hash_get (ngm->vtep4, ip41->dst_address.as_u32))
975                     goto exit1; /* no local VTEP for VXLAN packet */
976                   addr4 = ip41->dst_address;
977                 }
978             }
979           else
980             {
981               if (!ip6_address_is_equal (&addr6, &ip61->dst_address))
982                 {
983                   if (!hash_get_mem (ngm->vtep6, &ip61->dst_address))
984                     goto exit1; /* no local VTEP for VXLAN packet */
985                   addr6 = ip61->dst_address;
986                 }
987             }
988
989           flags1 = b1->flags;
990           good_udp1 = (flags1 & VNET_BUFFER_F_L4_CHECKSUM_CORRECT) != 0;
991
992           /* Don't verify UDP checksum for packets with explicit zero checksum. */
993           good_udp1 |= udp1->checksum == 0;
994
995           /* Verify UDP length */
996           if (is_ip4)
997             ip_len1 = clib_net_to_host_u16 (ip41->length);
998           else
999             ip_len1 = clib_net_to_host_u16 (ip61->payload_length);
1000           udp_len1 = clib_net_to_host_u16 (udp1->length);
1001           len_diff1 = ip_len1 - udp_len1;
1002
1003           /* Verify UDP checksum */
1004           if (PREDICT_FALSE (!good_udp1))
1005             {
1006               if ((flags1 & VNET_BUFFER_F_L4_CHECKSUM_COMPUTED) == 0)
1007                 {
1008                   if (is_ip4)
1009                     flags1 = ip4_tcp_udp_validate_checksum (vm, b1);
1010                   else
1011                     flags1 = ip6_tcp_udp_icmp_validate_checksum (vm, b1);
1012                   good_udp1 =
1013                     (flags1 & VNET_BUFFER_F_L4_CHECKSUM_CORRECT) != 0;
1014                 }
1015             }
1016
1017           if (is_ip4)
1018             {
1019               error1 = good_udp1 ? 0 : IP4_ERROR_UDP_CHECKSUM;
1020               error1 = (len_diff1 >= 0) ? error1 : IP4_ERROR_UDP_LENGTH;
1021             }
1022           else
1023             {
1024               error1 = good_udp1 ? 0 : IP6_ERROR_UDP_CHECKSUM;
1025               error1 = (len_diff1 >= 0) ? error1 : IP6_ERROR_UDP_LENGTH;
1026             }
1027
1028           next1 = error1 ?
1029             IP_VXLAN_BYPASS_NEXT_DROP : IP_VXLAN_BYPASS_NEXT_VXLAN;
1030           b1->error = error1 ? error_node->errors[error1] : 0;
1031
1032           /* vxlan_gpe-input node expect current at VXLAN header */
1033           if (is_ip4)
1034             vlib_buffer_advance (b1,
1035                                  sizeof (ip4_header_t) +
1036                                  sizeof (udp_header_t));
1037           else
1038             vlib_buffer_advance (b1,
1039                                  sizeof (ip6_header_t) +
1040                                  sizeof (udp_header_t));
1041
1042         exit1:
1043           vlib_validate_buffer_enqueue_x2 (vm, node, next_index,
1044                                            to_next, n_left_to_next,
1045                                            bi0, bi1, next0, next1);
1046         }
1047
1048       while (n_left_from > 0 && n_left_to_next > 0)
1049         {
1050           vlib_buffer_t *b0;
1051           ip4_header_t *ip40;
1052           ip6_header_t *ip60;
1053           udp_header_t *udp0;
1054           u32 bi0, ip_len0, udp_len0, flags0, next0;
1055           i32 len_diff0;
1056           u8 error0, good_udp0, proto0;
1057
1058           bi0 = to_next[0] = from[0];
1059           from += 1;
1060           n_left_from -= 1;
1061           to_next += 1;
1062           n_left_to_next -= 1;
1063
1064           b0 = vlib_get_buffer (vm, bi0);
1065           if (is_ip4)
1066             ip40 = vlib_buffer_get_current (b0);
1067           else
1068             ip60 = vlib_buffer_get_current (b0);
1069
1070           /* Setup packet for next IP feature */
1071           vnet_feature_next (&next0, b0);
1072
1073           if (is_ip4)
1074             proto0 = ip40->protocol;
1075           else
1076             proto0 = ip60->protocol;
1077
1078           if (proto0 != IP_PROTOCOL_UDP)
1079             goto exit;          /* not UDP packet */
1080
1081           if (is_ip4)
1082             udp0 = ip4_next_header (ip40);
1083           else
1084             udp0 = ip6_next_header (ip60);
1085
1086           if (udp0->dst_port != clib_host_to_net_u16 (UDP_DST_PORT_VXLAN_GPE))
1087             goto exit;          /* not VXLAN packet */
1088
1089           /* Validate DIP against VTEPs */
1090           if (is_ip4)
1091             {
1092               if (addr4.as_u32 != ip40->dst_address.as_u32)
1093                 {
1094                   if (!hash_get (ngm->vtep4, ip40->dst_address.as_u32))
1095                     goto exit;  /* no local VTEP for VXLAN packet */
1096                   addr4 = ip40->dst_address;
1097                 }
1098             }
1099           else
1100             {
1101               if (!ip6_address_is_equal (&addr6, &ip60->dst_address))
1102                 {
1103                   if (!hash_get_mem (ngm->vtep6, &ip60->dst_address))
1104                     goto exit;  /* no local VTEP for VXLAN packet */
1105                   addr6 = ip60->dst_address;
1106                 }
1107             }
1108
1109           flags0 = b0->flags;
1110           good_udp0 = (flags0 & VNET_BUFFER_F_L4_CHECKSUM_CORRECT) != 0;
1111
1112           /* Don't verify UDP checksum for packets with explicit zero checksum. */
1113           good_udp0 |= udp0->checksum == 0;
1114
1115           /* Verify UDP length */
1116           if (is_ip4)
1117             ip_len0 = clib_net_to_host_u16 (ip40->length);
1118           else
1119             ip_len0 = clib_net_to_host_u16 (ip60->payload_length);
1120           udp_len0 = clib_net_to_host_u16 (udp0->length);
1121           len_diff0 = ip_len0 - udp_len0;
1122
1123           /* Verify UDP checksum */
1124           if (PREDICT_FALSE (!good_udp0))
1125             {
1126               if ((flags0 & VNET_BUFFER_F_L4_CHECKSUM_COMPUTED) == 0)
1127                 {
1128                   if (is_ip4)
1129                     flags0 = ip4_tcp_udp_validate_checksum (vm, b0);
1130                   else
1131                     flags0 = ip6_tcp_udp_icmp_validate_checksum (vm, b0);
1132                   good_udp0 =
1133                     (flags0 & VNET_BUFFER_F_L4_CHECKSUM_CORRECT) != 0;
1134                 }
1135             }
1136
1137           if (is_ip4)
1138             {
1139               error0 = good_udp0 ? 0 : IP4_ERROR_UDP_CHECKSUM;
1140               error0 = (len_diff0 >= 0) ? error0 : IP4_ERROR_UDP_LENGTH;
1141             }
1142           else
1143             {
1144               error0 = good_udp0 ? 0 : IP6_ERROR_UDP_CHECKSUM;
1145               error0 = (len_diff0 >= 0) ? error0 : IP6_ERROR_UDP_LENGTH;
1146             }
1147
1148           next0 = error0 ?
1149             IP_VXLAN_BYPASS_NEXT_DROP : IP_VXLAN_BYPASS_NEXT_VXLAN;
1150           b0->error = error0 ? error_node->errors[error0] : 0;
1151
1152           /* vxlan_gpe-input node expect current at VXLAN header */
1153           if (is_ip4)
1154             vlib_buffer_advance (b0,
1155                                  sizeof (ip4_header_t) +
1156                                  sizeof (udp_header_t));
1157           else
1158             vlib_buffer_advance (b0,
1159                                  sizeof (ip6_header_t) +
1160                                  sizeof (udp_header_t));
1161
1162         exit:
1163           vlib_validate_buffer_enqueue_x1 (vm, node, next_index,
1164                                            to_next, n_left_to_next,
1165                                            bi0, next0);
1166         }
1167
1168       vlib_put_next_frame (vm, node, next_index, n_left_to_next);
1169     }
1170
1171   return frame->n_vectors;
1172 }
1173
1174 VLIB_NODE_FN (ip4_vxlan_gpe_bypass_node) (vlib_main_t * vm,
1175                                           vlib_node_runtime_t * node,
1176                                           vlib_frame_t * frame)
1177 {
1178   return ip_vxlan_gpe_bypass_inline (vm, node, frame, /* is_ip4 */ 1);
1179 }
1180
1181 /* *INDENT-OFF* */
1182 VLIB_REGISTER_NODE (ip4_vxlan_gpe_bypass_node) = {
1183   .name = "ip4-vxlan-gpe-bypass",
1184   .vector_size = sizeof (u32),
1185
1186   .n_next_nodes = IP_VXLAN_BYPASS_N_NEXT,
1187   .next_nodes = {
1188     [IP_VXLAN_BYPASS_NEXT_DROP] = "error-drop",
1189     [IP_VXLAN_BYPASS_NEXT_VXLAN] = "vxlan4-gpe-input",
1190   },
1191
1192   .format_buffer = format_ip4_header,
1193   .format_trace = format_ip4_forward_next_trace,
1194 };
1195 /* *INDENT-ON* */
1196
1197 #ifndef CLIB_MARCH_VARIANT
1198 /* Dummy init function to get us linked in. */
1199 clib_error_t *
1200 ip4_vxlan_gpe_bypass_init (vlib_main_t * vm)
1201 {
1202   return 0;
1203 }
1204
1205 VLIB_INIT_FUNCTION (ip4_vxlan_gpe_bypass_init);
1206 #endif /* CLIB_MARCH_VARIANT */
1207
1208 VLIB_NODE_FN (ip6_vxlan_gpe_bypass_node) (vlib_main_t * vm,
1209                                           vlib_node_runtime_t * node,
1210                                           vlib_frame_t * frame)
1211 {
1212   return ip_vxlan_gpe_bypass_inline (vm, node, frame, /* is_ip4 */ 0);
1213 }
1214
1215 /* *INDENT-OFF* */
1216 VLIB_REGISTER_NODE (ip6_vxlan_gpe_bypass_node) = {
1217   .name = "ip6-vxlan-gpe-bypass",
1218   .vector_size = sizeof (u32),
1219
1220   .n_next_nodes = IP_VXLAN_BYPASS_N_NEXT,
1221   .next_nodes = {
1222     [IP_VXLAN_BYPASS_NEXT_DROP] = "error-drop",
1223     [IP_VXLAN_BYPASS_NEXT_VXLAN] = "vxlan6-gpe-input",
1224   },
1225
1226   .format_buffer = format_ip6_header,
1227   .format_trace = format_ip6_forward_next_trace,
1228 };
1229 /* *INDENT-ON* */
1230
1231 #ifndef CLIB_MARCH_VARIANT
1232 /* Dummy init function to get us linked in. */
1233 clib_error_t *
1234 ip6_vxlan_gpe_bypass_init (vlib_main_t * vm)
1235 {
1236   return 0;
1237 }
1238
1239 VLIB_INIT_FUNCTION (ip6_vxlan_gpe_bypass_init);
1240 #endif /* CLIB_MARCH_VARIANT */
1241
1242 /*
1243  * fd.io coding-style-patch-verification: ON
1244  *
1245  * Local Variables:
1246  * eval: (c-set-style "gnu")
1247  * End:
1248  */