New upstream version 18.11.1
[deb_dpdk.git] / examples / ipsec-secgw / ipsec-secgw.c
index b45b87b..f88fdb4 100644 (file)
@@ -54,7 +54,7 @@
 #define NB_MBUF        (32000)
 
 #define CDEV_QUEUE_DESC 2048
-#define CDEV_MAP_ENTRIES 1024
+#define CDEV_MAP_ENTRIES 16384
 #define CDEV_MP_NB_OBJS 2048
 #define CDEV_MP_CACHE_SZ 64
 #define MAX_QUEUE_PAIRS 1
@@ -197,8 +197,7 @@ static struct rte_eth_conf port_conf = {
                .mq_mode        = ETH_MQ_RX_RSS,
                .max_rx_pkt_len = ETHER_MAX_LEN,
                .split_hdr_size = 0,
-               .offloads = DEV_RX_OFFLOAD_CHECKSUM |
-                           DEV_RX_OFFLOAD_CRC_STRIP,
+               .offloads = DEV_RX_OFFLOAD_CHECKSUM,
        },
        .rx_adv_conf = {
                .rss_conf = {
@@ -452,38 +451,55 @@ inbound_sp_sa(struct sp_ctx *sp, struct sa_ctx *sa, struct traffic_type *ip,
        ip->num = j;
 }
 
-static inline void
-process_pkts_inbound(struct ipsec_ctx *ipsec_ctx,
-               struct ipsec_traffic *traffic)
+static void
+split46_traffic(struct ipsec_traffic *trf, struct rte_mbuf *mb[], uint32_t num)
 {
+       uint32_t i, n4, n6;
+       struct ip *ip;
        struct rte_mbuf *m;
-       uint16_t idx, nb_pkts_in, i, n_ip4, n_ip6;
 
-       nb_pkts_in = ipsec_inbound(ipsec_ctx, traffic->ipsec.pkts,
-                       traffic->ipsec.num, MAX_PKT_BURST);
+       n4 = trf->ip4.num;
+       n6 = trf->ip6.num;
 
-       n_ip4 = traffic->ip4.num;
-       n_ip6 = traffic->ip6.num;
+       for (i = 0; i < num; i++) {
+
+               m = mb[i];
+               ip = rte_pktmbuf_mtod(m, struct ip *);
 
-       /* SP/ACL Inbound check ipsec and ip4 */
-       for (i = 0; i < nb_pkts_in; i++) {
-               m = traffic->ipsec.pkts[i];
-               struct ip *ip = rte_pktmbuf_mtod(m, struct ip *);
                if (ip->ip_v == IPVERSION) {
-                       idx = traffic->ip4.num++;
-                       traffic->ip4.pkts[idx] = m;
-                       traffic->ip4.data[idx] = rte_pktmbuf_mtod_offset(m,
+                       trf->ip4.pkts[n4] = m;
+                       trf->ip4.data[n4] = rte_pktmbuf_mtod_offset(m,
                                        uint8_t *, offsetof(struct ip, ip_p));
+                       n4++;
                } else if (ip->ip_v == IP6_VERSION) {
-                       idx = traffic->ip6.num++;
-                       traffic->ip6.pkts[idx] = m;
-                       traffic->ip6.data[idx] = rte_pktmbuf_mtod_offset(m,
+                       trf->ip6.pkts[n6] = m;
+                       trf->ip6.data[n6] = rte_pktmbuf_mtod_offset(m,
                                        uint8_t *,
                                        offsetof(struct ip6_hdr, ip6_nxt));
+                       n6++;
                } else
                        rte_pktmbuf_free(m);
        }
 
+       trf->ip4.num = n4;
+       trf->ip6.num = n6;
+}
+
+
+static inline void
+process_pkts_inbound(struct ipsec_ctx *ipsec_ctx,
+               struct ipsec_traffic *traffic)
+{
+       uint16_t nb_pkts_in, n_ip4, n_ip6;
+
+       n_ip4 = traffic->ip4.num;
+       n_ip6 = traffic->ip6.num;
+
+       nb_pkts_in = ipsec_inbound(ipsec_ctx, traffic->ipsec.pkts,
+                       traffic->ipsec.num, MAX_PKT_BURST);
+
+       split46_traffic(traffic, traffic->ipsec.pkts, nb_pkts_in);
+
        inbound_sp_sa(ipsec_ctx->sp4_ctx, ipsec_ctx->sa_ctx, &traffic->ip4,
                        n_ip4);
 
@@ -595,32 +611,45 @@ process_pkts_outbound_nosp(struct ipsec_ctx *ipsec_ctx,
                struct ipsec_traffic *traffic)
 {
        struct rte_mbuf *m;
-       uint32_t nb_pkts_out, i;
+       uint32_t nb_pkts_out, i, n;
        struct ip *ip;
 
        /* Drop any IPsec traffic from protected ports */
        for (i = 0; i < traffic->ipsec.num; i++)
                rte_pktmbuf_free(traffic->ipsec.pkts[i]);
 
-       traffic->ipsec.num = 0;
+       n = 0;
 
-       for (i = 0; i < traffic->ip4.num; i++)
-               traffic->ip4.res[i] = single_sa_idx;
+       for (i = 0; i < traffic->ip4.num; i++) {
+               traffic->ipsec.pkts[n] = traffic->ip4.pkts[i];
+               traffic->ipsec.res[n++] = single_sa_idx;
+       }
 
-       for (i = 0; i < traffic->ip6.num; i++)
-               traffic->ip6.res[i] = single_sa_idx;
+       for (i = 0; i < traffic->ip6.num; i++) {
+               traffic->ipsec.pkts[n] = traffic->ip6.pkts[i];
+               traffic->ipsec.res[n++] = single_sa_idx;
+       }
 
-       nb_pkts_out = ipsec_outbound(ipsec_ctx, traffic->ip4.pkts,
-                       traffic->ip4.res, traffic->ip4.num,
+       traffic->ip4.num = 0;
+       traffic->ip6.num = 0;
+       traffic->ipsec.num = n;
+
+       nb_pkts_out = ipsec_outbound(ipsec_ctx, traffic->ipsec.pkts,
+                       traffic->ipsec.res, traffic->ipsec.num,
                        MAX_PKT_BURST);
 
        /* They all sue the same SA (ip4 or ip6 tunnel) */
        m = traffic->ipsec.pkts[i];
        ip = rte_pktmbuf_mtod(m, struct ip *);
-       if (ip->ip_v == IPVERSION)
+       if (ip->ip_v == IPVERSION) {
                traffic->ip4.num = nb_pkts_out;
-       else
+               for (i = 0; i < nb_pkts_out; i++)
+                       traffic->ip4.pkts[i] = traffic->ipsec.pkts[i];
+       } else {
                traffic->ip6.num = nb_pkts_out;
+               for (i = 0; i < nb_pkts_out; i++)
+                       traffic->ip6.pkts[i] = traffic->ipsec.pkts[i];
+       }
 }
 
 static inline int32_t
@@ -778,7 +807,7 @@ process_pkts(struct lcore_conf *qconf, struct rte_mbuf **pkts,
 }
 
 static inline void
-drain_buffers(struct lcore_conf *qconf)
+drain_tx_buffers(struct lcore_conf *qconf)
 {
        struct buffer *buf;
        uint32_t portid;
@@ -792,6 +821,81 @@ drain_buffers(struct lcore_conf *qconf)
        }
 }
 
+static inline void
+drain_crypto_buffers(struct lcore_conf *qconf)
+{
+       uint32_t i;
+       struct ipsec_ctx *ctx;
+
+       /* drain inbound buffers*/
+       ctx = &qconf->inbound;
+       for (i = 0; i != ctx->nb_qps; i++) {
+               if (ctx->tbl[i].len != 0)
+                       enqueue_cop_burst(ctx->tbl  + i);
+       }
+
+       /* drain outbound buffers*/
+       ctx = &qconf->outbound;
+       for (i = 0; i != ctx->nb_qps; i++) {
+               if (ctx->tbl[i].len != 0)
+                       enqueue_cop_burst(ctx->tbl  + i);
+       }
+}
+
+static void
+drain_inbound_crypto_queues(const struct lcore_conf *qconf,
+               struct ipsec_ctx *ctx)
+{
+       uint32_t n;
+       struct ipsec_traffic trf;
+
+       /* dequeue packets from crypto-queue */
+       n = ipsec_inbound_cqp_dequeue(ctx, trf.ipsec.pkts,
+                       RTE_DIM(trf.ipsec.pkts));
+       if (n == 0)
+               return;
+
+       trf.ip4.num = 0;
+       trf.ip6.num = 0;
+
+       /* split traffic by ipv4-ipv6 */
+       split46_traffic(&trf, trf.ipsec.pkts, n);
+
+       /* process ipv4 packets */
+       inbound_sp_sa(ctx->sp4_ctx, ctx->sa_ctx, &trf.ip4, 0);
+       route4_pkts(qconf->rt4_ctx, trf.ip4.pkts, trf.ip4.num);
+
+       /* process ipv6 packets */
+       inbound_sp_sa(ctx->sp6_ctx, ctx->sa_ctx, &trf.ip6, 0);
+       route6_pkts(qconf->rt6_ctx, trf.ip6.pkts, trf.ip6.num);
+}
+
+static void
+drain_outbound_crypto_queues(const struct lcore_conf *qconf,
+               struct ipsec_ctx *ctx)
+{
+       uint32_t n;
+       struct ipsec_traffic trf;
+
+       /* dequeue packets from crypto-queue */
+       n = ipsec_outbound_cqp_dequeue(ctx, trf.ipsec.pkts,
+                       RTE_DIM(trf.ipsec.pkts));
+       if (n == 0)
+               return;
+
+       trf.ip4.num = 0;
+       trf.ip6.num = 0;
+
+       /* split traffic by ipv4-ipv6 */
+       split46_traffic(&trf, trf.ipsec.pkts, n);
+
+       /* process ipv4 packets */
+       route4_pkts(qconf->rt4_ctx, trf.ip4.pkts, trf.ip4.num);
+
+       /* process ipv6 packets */
+       route6_pkts(qconf->rt6_ctx, trf.ip6.pkts, trf.ip6.num);
+}
+
 /* main processing loop */
 static int32_t
 main_loop(__attribute__((unused)) void *dummy)
@@ -849,12 +953,14 @@ main_loop(__attribute__((unused)) void *dummy)
                diff_tsc = cur_tsc - prev_tsc;
 
                if (unlikely(diff_tsc > drain_tsc)) {
-                       drain_buffers(qconf);
+                       drain_tx_buffers(qconf);
+                       drain_crypto_buffers(qconf);
                        prev_tsc = cur_tsc;
                }
 
-               /* Read packet from RX queues */
                for (i = 0; i < qconf->nb_rx_queue; ++i) {
+
+                       /* Read packets from RX queues */
                        portid = rxql[i].port_id;
                        queueid = rxql[i].queue_id;
                        nb_rx = rte_eth_rx_burst(portid, queueid,
@@ -862,6 +968,14 @@ main_loop(__attribute__((unused)) void *dummy)
 
                        if (nb_rx > 0)
                                process_pkts(qconf, pkts, nb_rx, portid);
+
+                       /* dequeue and process completed crypto-ops */
+                       if (UNPROTECTED_PORT(portid))
+                               drain_inbound_crypto_queues(qconf,
+                                       &qconf->inbound);
+                       else
+                               drain_outbound_crypto_queues(qconf,
+                                       &qconf->outbound);
                }
        }
 }
@@ -1392,9 +1506,27 @@ cryptodevs_init(void)
 
        uint32_t max_sess_sz = 0, sess_sz;
        for (cdev_id = 0; cdev_id < rte_cryptodev_count(); cdev_id++) {
+               void *sec_ctx;
+
+               /* Get crypto priv session size */
                sess_sz = rte_cryptodev_sym_get_private_session_size(cdev_id);
                if (sess_sz > max_sess_sz)
                        max_sess_sz = sess_sz;
+
+               /*
+                * If crypto device is security capable, need to check the
+                * size of security session as well.
+                */
+
+               /* Get security context of the crypto device */
+               sec_ctx = rte_cryptodev_get_sec_ctx(cdev_id);
+               if (sec_ctx == NULL)
+                       continue;
+
+               /* Get size of security session */
+               sess_sz = rte_security_session_get_size(sec_ctx);
+               if (sess_sz > max_sess_sz)
+                       max_sess_sz = sess_sz;
        }
        RTE_ETH_FOREACH_DEV(port_id) {
                void *sec_ctx;