ipsec: IPSec interface correct drop w/ no protection
[vpp.git] / src / vnet / ipsec / ipsec_output.c
index 4056126..84927de 100644 (file)
@@ -20,8 +20,7 @@
 #include <vnet/ip/ip.h>
 
 #include <vnet/ipsec/ipsec.h>
-
-#if WITH_LIBSSL > 0
+#include <vnet/ipsec/ipsec_io.h>
 
 #define foreach_ipsec_output_error                   \
  _(RX_PKTS, "IPSec pkts received")                   \
@@ -64,9 +63,90 @@ format_ipsec_output_trace (u8 * s, va_list * args)
   return s;
 }
 
+always_inline void
+ipsec4_out_spd_add_flow_cache_entry (ipsec_main_t *im, u8 pr, u32 la, u32 ra,
+                                    u16 lp, u16 rp, u32 pol_id)
+{
+  u64 hash;
+  u8 overwrite = 0, stale_overwrite = 0;
+  ipsec4_spd_5tuple_t ip4_5tuple = { .ip4_addr = { (ip4_address_t) la,
+                                                  (ip4_address_t) ra },
+                                    .port = { lp, rp },
+                                    .proto = pr };
+
+  ip4_5tuple.kv_16_8.value = (((u64) pol_id) << 32) | ((u64) im->epoch_count);
+
+  hash = ipsec4_hash_16_8 (&ip4_5tuple.kv_16_8);
+  hash &= (im->ipsec4_out_spd_hash_num_buckets - 1);
+
+  ipsec_spinlock_lock (&im->ipsec4_out_spd_hash_tbl[hash].bucket_lock);
+  /* Check if we are overwriting an existing entry so we know
+  whether to increment the flow cache counter. Since flow
+  cache counter is reset on any policy add/remove, but
+  hash table values are not, we also need to check if the entry
+  we are overwriting is stale or not. If it's a stale entry
+  overwrite, we still want to increment flow cache counter */
+  overwrite = (im->ipsec4_out_spd_hash_tbl[hash].value != 0);
+  /* Check for stale entry by comparing with current epoch count */
+  if (PREDICT_FALSE (overwrite))
+    stale_overwrite =
+      (im->epoch_count !=
+       ((u32) (im->ipsec4_out_spd_hash_tbl[hash].value & 0xFFFFFFFF)));
+  clib_memcpy_fast (&im->ipsec4_out_spd_hash_tbl[hash], &ip4_5tuple.kv_16_8,
+                   sizeof (ip4_5tuple.kv_16_8));
+  ipsec_spinlock_unlock (&im->ipsec4_out_spd_hash_tbl[hash].bucket_lock);
+
+  /* Increment the counter to track active flow cache entries
+    when entering a fresh entry or overwriting a stale one */
+  if (!overwrite || stale_overwrite)
+    clib_atomic_fetch_add_relax (&im->ipsec4_out_spd_flow_cache_entries, 1);
+
+  return;
+}
+
 always_inline ipsec_policy_t *
-ipsec_output_policy_match (ipsec_spd_t * spd, u8 pr, u32 la, u32 ra, u16 lp,
-                          u16 rp)
+ipsec4_out_spd_find_flow_cache_entry (ipsec_main_t *im, u8 pr, u32 la, u32 ra,
+                                     u16 lp, u16 rp)
+{
+  ipsec_policy_t *p = NULL;
+  ipsec4_hash_kv_16_8_t kv_result;
+  u64 hash;
+
+  if (PREDICT_FALSE ((pr != IP_PROTOCOL_TCP) && (pr != IP_PROTOCOL_UDP) &&
+                    (pr != IP_PROTOCOL_SCTP)))
+    {
+      lp = 0;
+      rp = 0;
+    }
+  ipsec4_spd_5tuple_t ip4_5tuple = { .ip4_addr = { (ip4_address_t) la,
+                                                  (ip4_address_t) ra },
+                                    .port = { lp, rp },
+                                    .proto = pr };
+
+  hash = ipsec4_hash_16_8 (&ip4_5tuple.kv_16_8);
+  hash &= (im->ipsec4_out_spd_hash_num_buckets - 1);
+
+  ipsec_spinlock_lock (&im->ipsec4_out_spd_hash_tbl[hash].bucket_lock);
+  kv_result = im->ipsec4_out_spd_hash_tbl[hash];
+  ipsec_spinlock_unlock (&im->ipsec4_out_spd_hash_tbl[hash].bucket_lock);
+
+  if (ipsec4_hash_key_compare_16_8 ((u64 *) &ip4_5tuple.kv_16_8,
+                                   (u64 *) &kv_result))
+    {
+      if (im->epoch_count == ((u32) (kv_result.value & 0xFFFFFFFF)))
+       {
+         /* Get the policy based on the index */
+         p =
+           pool_elt_at_index (im->policies, ((u32) (kv_result.value >> 32)));
+       }
+    }
+
+  return p;
+}
+
+always_inline ipsec_policy_t *
+ipsec_output_policy_match (ipsec_spd_t *spd, u8 pr, u32 la, u32 ra, u16 lp,
+                          u16 rp, u8 flow_cache_enabled)
 {
   ipsec_main_t *im = &ipsec_main;
   ipsec_policy_t *p;
@@ -93,10 +173,13 @@ ipsec_output_policy_match (ipsec_spd_t * spd, u8 pr, u32 la, u32 ra, u16 lp,
     if (la > clib_net_to_host_u32 (p->laddr.stop.ip4.as_u32))
       continue;
 
-    if (PREDICT_FALSE
-       ((pr != IP_PROTOCOL_TCP) && (pr != IP_PROTOCOL_UDP)
-        && (pr != IP_PROTOCOL_SCTP)))
-      return p;
+    if (PREDICT_FALSE ((pr != IP_PROTOCOL_TCP) && (pr != IP_PROTOCOL_UDP) &&
+                      (pr != IP_PROTOCOL_SCTP)))
+      {
+       lp = 0;
+       rp = 0;
+       goto add_flow_cache;
+      }
 
     if (lp < p->lport.start)
       continue;
@@ -110,6 +193,15 @@ ipsec_output_policy_match (ipsec_spd_t * spd, u8 pr, u32 la, u32 ra, u16 lp,
     if (rp > p->rport.stop)
       continue;
 
+  add_flow_cache:
+    if (flow_cache_enabled)
+      {
+       /* Add an Entry in Flow cache */
+       ipsec4_out_spd_add_flow_cache_entry (
+         im, pr, clib_host_to_net_u32 (la), clib_host_to_net_u32 (ra),
+         clib_host_to_net_u16 (lp), clib_host_to_net_u16 (rp), *i);
+      }
+
     return p;
   }
   return 0;
@@ -186,6 +278,7 @@ ipsec_output_inline (vlib_main_t * vm, vlib_node_runtime_t * node,
   ipsec_spd_t *spd0 = 0;
   int bogus;
   u64 nc_protect = 0, nc_bypass = 0, nc_discard = 0, nc_nomatch = 0;
+  u8 flow_cache_enabled = im->flow_cache_flag;
 
   from = vlib_frame_vector_args (from_frame);
   n_left_from = from_frame->n_vectors;
@@ -193,9 +286,9 @@ ipsec_output_inline (vlib_main_t * vm, vlib_node_runtime_t * node,
 
   while (n_left_from > 0)
     {
-      u32 bi0, pi0;
-      vlib_buffer_t *b0;
-      ipsec_policy_t *p0;
+      u32 bi0, pi0, bi1;
+      vlib_buffer_t *b0, *b1;
+      ipsec_policy_t *p0 = NULL;
       ip4_header_t *ip0;
       ip6_header_t *ip6_0 = 0;
       udp_header_t *udp0;
@@ -205,6 +298,13 @@ ipsec_output_inline (vlib_main_t * vm, vlib_node_runtime_t * node,
 
       bi0 = from[0];
       b0 = vlib_get_buffer (vm, bi0);
+      if (n_left_from > 1)
+       {
+         bi1 = from[1];
+         b1 = vlib_get_buffer (vm, bi1);
+         CLIB_PREFETCH (b1, CLIB_CACHE_LINE_BYTES * 2, STORE);
+         vlib_prefetch_buffer_data (b1, LOAD);
+       }
       sw_if_index0 = vnet_buffer (b0)->sw_if_index[VLIB_TX];
       iph_offset = vnet_buffer (b0)->ip.save_rewrite_length;
       ip0 = (ip4_header_t *) ((u8 *) vlib_buffer_get_current (b0)
@@ -214,7 +314,7 @@ ipsec_output_inline (vlib_main_t * vm, vlib_node_runtime_t * node,
       if (PREDICT_FALSE (last_sw_if_index != sw_if_index0))
        {
          uword *p = hash_get (im->spd_index_by_sw_if_index, sw_if_index0);
-         ASSERT (p);
+         ALWAYS_ASSERT (p);
          spd_index0 = p[0];
          spd0 = pool_elt_at_index (im->spds, spd_index0);
          last_sw_if_index = sw_if_index0;
@@ -256,15 +356,26 @@ ipsec_output_inline (vlib_main_t * vm, vlib_node_runtime_t * node,
                        sw_if_index0, spd_index0, spd0->id);
 #endif
 
-         p0 = ipsec_output_policy_match (spd0, ip0->protocol,
-                                         clib_net_to_host_u32
-                                         (ip0->src_address.as_u32),
-                                         clib_net_to_host_u32
-                                         (ip0->dst_address.as_u32),
-                                         clib_net_to_host_u16
-                                         (udp0->src_port),
-                                         clib_net_to_host_u16
-                                         (udp0->dst_port));
+         /*
+          * Check whether flow cache is enabled.
+          */
+         if (flow_cache_enabled)
+           {
+             p0 = ipsec4_out_spd_find_flow_cache_entry (
+               im, ip0->protocol, ip0->src_address.as_u32,
+               ip0->dst_address.as_u32, udp0->src_port, udp0->dst_port);
+           }
+
+         /* Fall back to linear search if flow cache lookup fails */
+         if (p0 == NULL)
+           {
+             p0 = ipsec_output_policy_match (
+               spd0, ip0->protocol,
+               clib_net_to_host_u32 (ip0->src_address.as_u32),
+               clib_net_to_host_u32 (ip0->dst_address.as_u32),
+               clib_net_to_host_u16 (udp0->src_port),
+               clib_net_to_host_u16 (udp0->dst_port), flow_cache_enabled);
+           }
        }
       tcp0 = (void *) udp0;
 
@@ -289,7 +400,7 @@ ipsec_output_inline (vlib_main_t * vm, vlib_node_runtime_t * node,
            {
              ipsec_sa_t *sa = 0;
              nc_protect++;
-             sa = pool_elt_at_index (im->sad, p0->sa_index);
+             sa = ipsec_sa_get (p0->sa_index);
              if (sa->protocol == IPSEC_PROTOCOL_ESP)
                if (is_ipv6)
                  next_node_index = im->esp6_encrypt_node_index;
@@ -301,45 +412,50 @@ ipsec_output_inline (vlib_main_t * vm, vlib_node_runtime_t * node,
                next_node_index = im->ah4_encrypt_node_index;
              vnet_buffer (b0)->ipsec.sad_index = p0->sa_index;
 
-             if (is_ipv6)
-               {
-                 if (PREDICT_FALSE
-                     (b0->flags & VNET_BUFFER_F_OFFLOAD_TCP_CKSUM))
-                   {
-                     tcp0->checksum =
-                       ip6_tcp_udp_icmp_compute_checksum (vm, b0, ip6_0,
-                                                          &bogus);
-                     b0->flags &= ~VNET_BUFFER_F_OFFLOAD_TCP_CKSUM;
-                   }
-                 if (PREDICT_FALSE
-                     (b0->flags & VNET_BUFFER_F_OFFLOAD_UDP_CKSUM))
-                   {
-                     udp0->checksum =
-                       ip6_tcp_udp_icmp_compute_checksum (vm, b0, ip6_0,
-                                                          &bogus);
-                     b0->flags &= ~VNET_BUFFER_F_OFFLOAD_UDP_CKSUM;
-                   }
-               }
-             else
+             if (PREDICT_FALSE (b0->flags & VNET_BUFFER_F_OFFLOAD))
                {
-                 if (b0->flags & VNET_BUFFER_F_OFFLOAD_IP_CKSUM)
-                   {
-                     ip0->checksum = ip4_header_checksum (ip0);
-                     b0->flags &= ~VNET_BUFFER_F_OFFLOAD_IP_CKSUM;
-                   }
-                 if (PREDICT_FALSE
-                     (b0->flags & VNET_BUFFER_F_OFFLOAD_TCP_CKSUM))
+                 vnet_buffer_oflags_t oflags = vnet_buffer (b0)->oflags;
+
+                 /*
+                  * Clearing offload flags before checksum is computed
+                  * It guarantees the cache hit!
+                  */
+                 vnet_buffer_offload_flags_clear (b0, oflags);
+
+                 if (is_ipv6)
                    {
-                     tcp0->checksum =
-                       ip4_tcp_udp_compute_checksum (vm, b0, ip0);
-                     b0->flags &= ~VNET_BUFFER_F_OFFLOAD_TCP_CKSUM;
+                     if (PREDICT_FALSE (oflags &
+                                        VNET_BUFFER_OFFLOAD_F_TCP_CKSUM))
+                       {
+                         tcp0->checksum = ip6_tcp_udp_icmp_compute_checksum (
+                           vm, b0, ip6_0, &bogus);
+                       }
+                     if (PREDICT_FALSE (oflags &
+                                        VNET_BUFFER_OFFLOAD_F_UDP_CKSUM))
+                       {
+                         udp0->checksum = ip6_tcp_udp_icmp_compute_checksum (
+                           vm, b0, ip6_0, &bogus);
+                       }
                    }
-                 if (PREDICT_FALSE
-                     (b0->flags & VNET_BUFFER_F_OFFLOAD_UDP_CKSUM))
+                 else
                    {
-                     udp0->checksum =
-                       ip4_tcp_udp_compute_checksum (vm, b0, ip0);
-                     b0->flags &= ~VNET_BUFFER_F_OFFLOAD_UDP_CKSUM;
+                     if (PREDICT_FALSE (oflags &
+                                        VNET_BUFFER_OFFLOAD_F_IP_CKSUM))
+                       {
+                         ip0->checksum = ip4_header_checksum (ip0);
+                       }
+                     if (PREDICT_FALSE (oflags &
+                                        VNET_BUFFER_OFFLOAD_F_TCP_CKSUM))
+                       {
+                         tcp0->checksum =
+                           ip4_tcp_udp_compute_checksum (vm, b0, ip0);
+                       }
+                     if (PREDICT_FALSE (oflags &
+                                        VNET_BUFFER_OFFLOAD_F_UDP_CKSUM))
+                       {
+                         udp0->checksum =
+                           ip4_tcp_udp_compute_checksum (vm, b0, ip0);
+                       }
                    }
                }
              vlib_buffer_advance (b0, iph_offset);
@@ -445,7 +561,6 @@ VLIB_NODE_FN (ipsec6_output_node) (vlib_main_t * vm,
   return ipsec_output_inline (vm, node, frame, 1);
 }
 
-/* *INDENT-OFF* */
 VLIB_REGISTER_NODE (ipsec6_output_node) = {
   .name = "ipsec6-output-feature",
   .vector_size = sizeof (u32),
@@ -462,39 +577,4 @@ VLIB_REGISTER_NODE (ipsec6_output_node) = {
 #undef _
   },
 };
-/* *INDENT-ON* */
-
-#else /* IPSEC > 1 */
-
-/* Dummy ipsec output node, in case when IPSec is disabled */
-
-static uword
-ipsec_output_node_fn (vlib_main_t * vm,
-                     vlib_node_runtime_t * node, vlib_frame_t * frame)
-{
-  clib_warning ("IPSec disabled");
-  return 0;
-}
 
-/* *INDENT-OFF* */
-VLIB_REGISTER_NODE (ipsec4_output_node) = {
-  .vector_size = sizeof (u32),
-  .function = ipsec_output_node_fn,
-  .name = "ipsec4-output-feature",
-};
-
-VLIB_REGISTER_NODE (ipsec6_output_node) = {
-  .vector_size = sizeof (u32),
-  .function = ipsec_output_node_fn,
-  .name = "ipsec6-output-feature",
-};
-/* *INDENT-ON* */
-#endif
-
-/*
- * fd.io coding-style-patch-verification: ON
- *
- * Local Variables:
- * eval: (c-set-style "gnu")
- * End:
- */