ipsec: move the IPSec SA pool out of ipsec_main
[vpp.git] / src / vnet / ipsec / ipsec_output.c
1 /*
2  * ipsec_output.c : IPSec output node
3  *
4  * Copyright (c) 2015 Cisco and/or its affiliates.
5  * Licensed under the Apache License, Version 2.0 (the "License");
6  * you may not use this file except in compliance with the License.
7  * You may obtain a copy of the License at:
8  *
9  *     http://www.apache.org/licenses/LICENSE-2.0
10  *
11  * Unless required by applicable law or agreed to in writing, software
12  * distributed under the License is distributed on an "AS IS" BASIS,
13  * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
14  * See the License for the specific language governing permissions and
15  * limitations under the License.
16  */
17
18 #include <vnet/vnet.h>
19 #include <vnet/api_errno.h>
20 #include <vnet/ip/ip.h>
21
22 #include <vnet/ipsec/ipsec.h>
23 #include <vnet/ipsec/ipsec_io.h>
24
25 #if WITH_LIBSSL > 0
26
27 #define foreach_ipsec_output_error                   \
28  _(RX_PKTS, "IPSec pkts received")                   \
29  _(POLICY_DISCARD, "IPSec policy discard")           \
30  _(POLICY_NO_MATCH, "IPSec policy (no match)")       \
31  _(POLICY_PROTECT, "IPSec policy protect")           \
32  _(POLICY_BYPASS, "IPSec policy bypass")             \
33  _(ENCAPS_FAILED, "IPSec encapsulation failed")
34
35 typedef enum
36 {
37 #define _(sym,str) IPSEC_OUTPUT_ERROR_##sym,
38   foreach_ipsec_output_error
39 #undef _
40     IPSEC_DECAP_N_ERROR,
41 } ipsec_output_error_t;
42
43 static char *ipsec_output_error_strings[] = {
44 #define _(sym,string) string,
45   foreach_ipsec_output_error
46 #undef _
47 };
48
49 typedef struct
50 {
51   u32 spd_id;
52   u32 policy_id;
53 } ipsec_output_trace_t;
54
55 /* packet trace format function */
56 static u8 *
57 format_ipsec_output_trace (u8 * s, va_list * args)
58 {
59   CLIB_UNUSED (vlib_main_t * vm) = va_arg (*args, vlib_main_t *);
60   CLIB_UNUSED (vlib_node_t * node) = va_arg (*args, vlib_node_t *);
61   ipsec_output_trace_t *t = va_arg (*args, ipsec_output_trace_t *);
62
63   s = format (s, "spd %u policy %d", t->spd_id, t->policy_id);
64
65   return s;
66 }
67
68 always_inline ipsec_policy_t *
69 ipsec_output_policy_match (ipsec_spd_t * spd, u8 pr, u32 la, u32 ra, u16 lp,
70                            u16 rp)
71 {
72   ipsec_main_t *im = &ipsec_main;
73   ipsec_policy_t *p;
74   u32 *i;
75
76   if (!spd)
77     return 0;
78
79   vec_foreach (i, spd->policies[IPSEC_SPD_POLICY_IP4_OUTBOUND])
80   {
81     p = pool_elt_at_index (im->policies, *i);
82     if (PREDICT_FALSE (p->protocol && (p->protocol != pr)))
83       continue;
84
85     if (ra < clib_net_to_host_u32 (p->raddr.start.ip4.as_u32))
86       continue;
87
88     if (ra > clib_net_to_host_u32 (p->raddr.stop.ip4.as_u32))
89       continue;
90
91     if (la < clib_net_to_host_u32 (p->laddr.start.ip4.as_u32))
92       continue;
93
94     if (la > clib_net_to_host_u32 (p->laddr.stop.ip4.as_u32))
95       continue;
96
97     if (PREDICT_FALSE
98         ((pr != IP_PROTOCOL_TCP) && (pr != IP_PROTOCOL_UDP)
99          && (pr != IP_PROTOCOL_SCTP)))
100       return p;
101
102     if (lp < p->lport.start)
103       continue;
104
105     if (lp > p->lport.stop)
106       continue;
107
108     if (rp < p->rport.start)
109       continue;
110
111     if (rp > p->rport.stop)
112       continue;
113
114     return p;
115   }
116   return 0;
117 }
118
119 always_inline uword
120 ip6_addr_match_range (ip6_address_t * a, ip6_address_t * la,
121                       ip6_address_t * ua)
122 {
123   if ((memcmp (a->as_u64, la->as_u64, 2 * sizeof (u64)) >= 0) &&
124       (memcmp (a->as_u64, ua->as_u64, 2 * sizeof (u64)) <= 0))
125     return 1;
126   return 0;
127 }
128
129 always_inline ipsec_policy_t *
130 ipsec6_output_policy_match (ipsec_spd_t * spd,
131                             ip6_address_t * la,
132                             ip6_address_t * ra, u16 lp, u16 rp, u8 pr)
133 {
134   ipsec_main_t *im = &ipsec_main;
135   ipsec_policy_t *p;
136   u32 *i;
137
138   if (!spd)
139     return 0;
140
141   vec_foreach (i, spd->policies[IPSEC_SPD_POLICY_IP6_OUTBOUND])
142   {
143     p = pool_elt_at_index (im->policies, *i);
144     if (PREDICT_FALSE (p->protocol && (p->protocol != pr)))
145       continue;
146
147     if (!ip6_addr_match_range (ra, &p->raddr.start.ip6, &p->raddr.stop.ip6))
148       continue;
149
150     if (!ip6_addr_match_range (la, &p->laddr.start.ip6, &p->laddr.stop.ip6))
151       continue;
152
153     if (PREDICT_FALSE
154         ((pr != IP_PROTOCOL_TCP) && (pr != IP_PROTOCOL_UDP)
155          && (pr != IP_PROTOCOL_SCTP)))
156       return p;
157
158     if (lp < p->lport.start)
159       continue;
160
161     if (lp > p->lport.stop)
162       continue;
163
164     if (rp < p->rport.start)
165       continue;
166
167     if (rp > p->rport.stop)
168       continue;
169
170     return p;
171   }
172
173   return 0;
174 }
175
176 static inline uword
177 ipsec_output_inline (vlib_main_t * vm, vlib_node_runtime_t * node,
178                      vlib_frame_t * from_frame, int is_ipv6)
179 {
180   ipsec_main_t *im = &ipsec_main;
181
182   u32 *from, *to_next = 0, thread_index;
183   u32 n_left_from, sw_if_index0, last_sw_if_index = (u32) ~ 0;
184   u32 next_node_index = (u32) ~ 0, last_next_node_index = (u32) ~ 0;
185   vlib_frame_t *f = 0;
186   u32 spd_index0 = ~0;
187   ipsec_spd_t *spd0 = 0;
188   int bogus;
189   u64 nc_protect = 0, nc_bypass = 0, nc_discard = 0, nc_nomatch = 0;
190
191   from = vlib_frame_vector_args (from_frame);
192   n_left_from = from_frame->n_vectors;
193   thread_index = vm->thread_index;
194
195   while (n_left_from > 0)
196     {
197       u32 bi0, pi0, bi1;
198       vlib_buffer_t *b0, *b1;
199       ipsec_policy_t *p0;
200       ip4_header_t *ip0;
201       ip6_header_t *ip6_0 = 0;
202       udp_header_t *udp0;
203       u32 iph_offset = 0;
204       tcp_header_t *tcp0;
205       u64 bytes0;
206
207       bi0 = from[0];
208       b0 = vlib_get_buffer (vm, bi0);
209       if (n_left_from > 1)
210         {
211           bi1 = from[1];
212           b1 = vlib_get_buffer (vm, bi1);
213           CLIB_PREFETCH (b1, CLIB_CACHE_LINE_BYTES * 2, STORE);
214           vlib_prefetch_buffer_data (b1, LOAD);
215         }
216       sw_if_index0 = vnet_buffer (b0)->sw_if_index[VLIB_TX];
217       iph_offset = vnet_buffer (b0)->ip.save_rewrite_length;
218       ip0 = (ip4_header_t *) ((u8 *) vlib_buffer_get_current (b0)
219                               + iph_offset);
220
221       /* lookup for SPD only if sw_if_index is changed */
222       if (PREDICT_FALSE (last_sw_if_index != sw_if_index0))
223         {
224           uword *p = hash_get (im->spd_index_by_sw_if_index, sw_if_index0);
225           ALWAYS_ASSERT (p);
226           spd_index0 = p[0];
227           spd0 = pool_elt_at_index (im->spds, spd_index0);
228           last_sw_if_index = sw_if_index0;
229         }
230
231       if (is_ipv6)
232         {
233           ip6_0 = (ip6_header_t *) ((u8 *) vlib_buffer_get_current (b0)
234                                     + iph_offset);
235
236           udp0 = ip6_next_header (ip6_0);
237 #if 0
238           clib_warning
239             ("packet received from %U port %u to %U port %u spd_id %u",
240              format_ip6_address, &ip6_0->src_address,
241              clib_net_to_host_u16 (udp0->src_port), format_ip6_address,
242              &ip6_0->dst_address, clib_net_to_host_u16 (udp0->dst_port),
243              spd0->id);
244 #endif
245
246           p0 = ipsec6_output_policy_match (spd0,
247                                            &ip6_0->src_address,
248                                            &ip6_0->dst_address,
249                                            clib_net_to_host_u16
250                                            (udp0->src_port),
251                                            clib_net_to_host_u16
252                                            (udp0->dst_port), ip6_0->protocol);
253         }
254       else
255         {
256           udp0 = (udp_header_t *) ((u8 *) ip0 + ip4_header_bytes (ip0));
257
258 #if 0
259           clib_warning ("packet received from %U to %U port %u",
260                         format_ip4_address, ip0->src_address.as_u8,
261                         format_ip4_address, ip0->dst_address.as_u8,
262                         clib_net_to_host_u16 (udp0->dst_port));
263           clib_warning ("sw_if_index0 %u spd_index0 %u spd_id %u",
264                         sw_if_index0, spd_index0, spd0->id);
265 #endif
266
267           p0 = ipsec_output_policy_match (spd0, ip0->protocol,
268                                           clib_net_to_host_u32
269                                           (ip0->src_address.as_u32),
270                                           clib_net_to_host_u32
271                                           (ip0->dst_address.as_u32),
272                                           clib_net_to_host_u16
273                                           (udp0->src_port),
274                                           clib_net_to_host_u16
275                                           (udp0->dst_port));
276         }
277       tcp0 = (void *) udp0;
278
279       if (PREDICT_TRUE (p0 != NULL))
280         {
281           pi0 = p0 - im->policies;
282
283           vlib_prefetch_combined_counter (&ipsec_spd_policy_counters,
284                                           thread_index, pi0);
285
286           if (is_ipv6)
287             {
288               bytes0 = clib_net_to_host_u16 (ip6_0->payload_length);
289               bytes0 += sizeof (ip6_header_t);
290             }
291           else
292             {
293               bytes0 = clib_net_to_host_u16 (ip0->length);
294             }
295
296           if (p0->policy == IPSEC_POLICY_ACTION_PROTECT)
297             {
298               ipsec_sa_t *sa = 0;
299               nc_protect++;
300               sa = ipsec_sa_get (p0->sa_index);
301               if (sa->protocol == IPSEC_PROTOCOL_ESP)
302                 if (is_ipv6)
303                   next_node_index = im->esp6_encrypt_node_index;
304                 else
305                   next_node_index = im->esp4_encrypt_node_index;
306               else if (is_ipv6)
307                 next_node_index = im->ah6_encrypt_node_index;
308               else
309                 next_node_index = im->ah4_encrypt_node_index;
310               vnet_buffer (b0)->ipsec.sad_index = p0->sa_index;
311
312               if (PREDICT_FALSE (b0->flags & VNET_BUFFER_F_OFFLOAD))
313                 {
314                   u32 oflags = vnet_buffer2 (b0)->oflags;
315
316                   /*
317                    * Clearing offload flags before checksum is computed
318                    * It guarantees the cache hit!
319                    */
320                   vnet_buffer_offload_flags_clear (b0, oflags);
321
322                   if (is_ipv6)
323                     {
324                       if (PREDICT_FALSE (oflags &
325                                          VNET_BUFFER_OFFLOAD_F_TCP_CKSUM))
326                         {
327                           tcp0->checksum = ip6_tcp_udp_icmp_compute_checksum (
328                             vm, b0, ip6_0, &bogus);
329                         }
330                       if (PREDICT_FALSE (oflags &
331                                          VNET_BUFFER_OFFLOAD_F_UDP_CKSUM))
332                         {
333                           udp0->checksum = ip6_tcp_udp_icmp_compute_checksum (
334                             vm, b0, ip6_0, &bogus);
335                         }
336                     }
337                   else
338                     {
339                       if (PREDICT_FALSE (oflags &
340                                          VNET_BUFFER_OFFLOAD_F_IP_CKSUM))
341                         {
342                           ip0->checksum = ip4_header_checksum (ip0);
343                         }
344                       if (PREDICT_FALSE (oflags &
345                                          VNET_BUFFER_OFFLOAD_F_TCP_CKSUM))
346                         {
347                           tcp0->checksum =
348                             ip4_tcp_udp_compute_checksum (vm, b0, ip0);
349                         }
350                       if (PREDICT_FALSE (oflags &
351                                          VNET_BUFFER_OFFLOAD_F_UDP_CKSUM))
352                         {
353                           udp0->checksum =
354                             ip4_tcp_udp_compute_checksum (vm, b0, ip0);
355                         }
356                     }
357                 }
358               vlib_buffer_advance (b0, iph_offset);
359             }
360           else if (p0->policy == IPSEC_POLICY_ACTION_BYPASS)
361             {
362               nc_bypass++;
363               next_node_index = get_next_output_feature_node_index (b0, node);
364             }
365           else
366             {
367               nc_discard++;
368               next_node_index = im->error_drop_node_index;
369             }
370           vlib_increment_combined_counter
371             (&ipsec_spd_policy_counters, thread_index, pi0, 1, bytes0);
372         }
373       else
374         {
375           pi0 = ~0;
376           nc_nomatch++;
377           next_node_index = im->error_drop_node_index;
378         }
379
380       from += 1;
381       n_left_from -= 1;
382
383       if (PREDICT_FALSE ((last_next_node_index != next_node_index) || f == 0))
384         {
385           /* if this is not 1st frame */
386           if (f)
387             vlib_put_frame_to_node (vm, last_next_node_index, f);
388
389           last_next_node_index = next_node_index;
390
391           f = vlib_get_frame_to_node (vm, next_node_index);
392
393           /* frame->frame_flags, copy it from node */
394           /* Copy trace flag from next_frame and from runtime. */
395           f->frame_flags |= node->flags & VLIB_NODE_FLAG_TRACE;
396
397           to_next = vlib_frame_vector_args (f);
398         }
399
400       to_next[0] = bi0;
401       to_next += 1;
402       f->n_vectors++;
403
404       if (PREDICT_FALSE (node->flags & VLIB_NODE_FLAG_TRACE) &&
405           PREDICT_FALSE (b0->flags & VLIB_BUFFER_IS_TRACED))
406         {
407           ipsec_output_trace_t *tr =
408             vlib_add_trace (vm, node, b0, sizeof (*tr));
409           if (spd0)
410             tr->spd_id = spd0->id;
411           tr->policy_id = pi0;
412         }
413     }
414
415   vlib_put_frame_to_node (vm, next_node_index, f);
416   vlib_node_increment_counter (vm, node->node_index,
417                                IPSEC_OUTPUT_ERROR_POLICY_PROTECT, nc_protect);
418   vlib_node_increment_counter (vm, node->node_index,
419                                IPSEC_OUTPUT_ERROR_POLICY_BYPASS, nc_bypass);
420   vlib_node_increment_counter (vm, node->node_index,
421                                IPSEC_OUTPUT_ERROR_POLICY_DISCARD, nc_discard);
422   vlib_node_increment_counter (vm, node->node_index,
423                                IPSEC_OUTPUT_ERROR_POLICY_NO_MATCH,
424                                nc_nomatch);
425   return from_frame->n_vectors;
426 }
427
428 VLIB_NODE_FN (ipsec4_output_node) (vlib_main_t * vm,
429                                    vlib_node_runtime_t * node,
430                                    vlib_frame_t * frame)
431 {
432   return ipsec_output_inline (vm, node, frame, 0);
433 }
434
435 /* *INDENT-OFF* */
436 VLIB_REGISTER_NODE (ipsec4_output_node) = {
437   .name = "ipsec4-output-feature",
438   .vector_size = sizeof (u32),
439   .format_trace = format_ipsec_output_trace,
440   .type = VLIB_NODE_TYPE_INTERNAL,
441
442   .n_errors = ARRAY_LEN(ipsec_output_error_strings),
443   .error_strings = ipsec_output_error_strings,
444
445   .n_next_nodes = IPSEC_OUTPUT_N_NEXT,
446   .next_nodes = {
447 #define _(s,n) [IPSEC_OUTPUT_NEXT_##s] = n,
448     foreach_ipsec_output_next
449 #undef _
450   },
451 };
452 /* *INDENT-ON* */
453
454 VLIB_NODE_FN (ipsec6_output_node) (vlib_main_t * vm,
455                                    vlib_node_runtime_t * node,
456                                    vlib_frame_t * frame)
457 {
458   return ipsec_output_inline (vm, node, frame, 1);
459 }
460
461 /* *INDENT-OFF* */
462 VLIB_REGISTER_NODE (ipsec6_output_node) = {
463   .name = "ipsec6-output-feature",
464   .vector_size = sizeof (u32),
465   .format_trace = format_ipsec_output_trace,
466   .type = VLIB_NODE_TYPE_INTERNAL,
467
468   .n_errors = ARRAY_LEN(ipsec_output_error_strings),
469   .error_strings = ipsec_output_error_strings,
470
471   .n_next_nodes = IPSEC_OUTPUT_N_NEXT,
472   .next_nodes = {
473 #define _(s,n) [IPSEC_OUTPUT_NEXT_##s] = n,
474     foreach_ipsec_output_next
475 #undef _
476   },
477 };
478 /* *INDENT-ON* */
479
480 #else /* IPSEC > 1 */
481
482 /* Dummy ipsec output node, in case when IPSec is disabled */
483
484 static uword
485 ipsec_output_node_fn (vlib_main_t * vm,
486                       vlib_node_runtime_t * node, vlib_frame_t * frame)
487 {
488   return 0;
489 }
490
491 /* *INDENT-OFF* */
492 VLIB_REGISTER_NODE (ipsec4_output_node) = {
493   .vector_size = sizeof (u32),
494   .function = ipsec_output_node_fn,
495   .name = "ipsec4-output-feature",
496 };
497
498 VLIB_REGISTER_NODE (ipsec6_output_node) = {
499   .vector_size = sizeof (u32),
500   .function = ipsec_output_node_fn,
501   .name = "ipsec6-output-feature",
502 };
503 /* *INDENT-ON* */
504 #endif
505
506 /*
507  * fd.io coding-style-patch-verification: ON
508  *
509  * Local Variables:
510  * eval: (c-set-style "gnu")
511  * End:
512  */