New upstream version 17.11-rc3
[deb_dpdk.git] / drivers / net / bnxt / bnxt_rxq.c
index 0793820..c4da474 100644 (file)
@@ -60,10 +60,13 @@ void bnxt_free_rxq_stats(struct bnxt_rx_queue *rxq)
 int bnxt_mq_rx_configure(struct bnxt *bp)
 {
        struct rte_eth_conf *dev_conf = &bp->eth_dev->data->dev_conf;
-       unsigned int i, j, nb_q_per_grp, ring_idx;
-       int start_grp_id, end_grp_id, rc = 0;
+       const struct rte_eth_vmdq_rx_conf *conf =
+                   &dev_conf->rx_adv_conf.vmdq_rx_conf;
+       unsigned int i, j, nb_q_per_grp = 1, ring_idx = 0;
+       int start_grp_id, end_grp_id = 1, rc = 0;
        struct bnxt_vnic_info *vnic;
        struct bnxt_filter_info *filter;
+       enum rte_eth_nb_pools pools = bp->rx_cp_nr_rings, max_pools = 0;
        struct bnxt_rx_queue *rxq;
 
        bp->nr_vnics = 0;
@@ -98,117 +101,125 @@ int bnxt_mq_rx_configure(struct bnxt *bp)
        }
 
        /* Multi-queue mode */
-       if (dev_conf->rxmode.mq_mode & ETH_MQ_RX_VMDQ_FLAG) {
+       if (dev_conf->rxmode.mq_mode & ETH_MQ_RX_VMDQ_DCB_RSS) {
                /* VMDq ONLY, VMDq+RSS, VMDq+DCB, VMDq+DCB+RSS */
-               enum rte_eth_nb_pools pools;
 
                switch (dev_conf->rxmode.mq_mode) {
                case ETH_MQ_RX_VMDQ_RSS:
                case ETH_MQ_RX_VMDQ_ONLY:
-                       {
-                               const struct rte_eth_vmdq_rx_conf *conf =
-                                   &dev_conf->rx_adv_conf.vmdq_rx_conf;
-
-                               /* ETH_8/64_POOLs */
-                               pools = conf->nb_queue_pools;
-                               break;
-                       }
+                       /* ETH_8/64_POOLs */
+                       pools = conf->nb_queue_pools;
+                       /* For each pool, allocate MACVLAN CFA rule & VNIC */
+                       max_pools = RTE_MIN(bp->max_vnics,
+                                           RTE_MIN(bp->max_l2_ctx,
+                                           RTE_MIN(bp->max_rsscos_ctx,
+                                                   ETH_64_POOLS)));
+                       if (pools > max_pools)
+                               pools = max_pools;
+                       break;
+               case ETH_MQ_RX_RSS:
+                       pools = bp->rx_cp_nr_rings;
+                       break;
                default:
                        RTE_LOG(ERR, PMD, "Unsupported mq_mod %d\n",
                                dev_conf->rxmode.mq_mode);
                        rc = -EINVAL;
                        goto err_out;
                }
-               /* For each pool, allocate MACVLAN CFA rule & VNIC */
-               if (!pools) {
-                       pools = RTE_MIN(bp->max_vnics,
-                           RTE_MIN(bp->max_l2_ctx,
-                            RTE_MIN(bp->max_rsscos_ctx, ETH_64_POOLS)));
-                       RTE_LOG(ERR, PMD,
-                               "VMDq pool not set, defaulted to 64\n");
-                       pools = ETH_64_POOLS;
+       }
+
+       nb_q_per_grp = bp->rx_cp_nr_rings / pools;
+       start_grp_id = 0;
+       end_grp_id = nb_q_per_grp;
+
+       for (i = 0; i < pools; i++) {
+               vnic = bnxt_alloc_vnic(bp);
+               if (!vnic) {
+                       RTE_LOG(ERR, PMD, "VNIC alloc failed\n");
+                       rc = -ENOMEM;
+                       goto err_out;
                }
-               nb_q_per_grp = bp->rx_cp_nr_rings / pools;
-               start_grp_id = 0;
-               end_grp_id = nb_q_per_grp;
-
-               ring_idx = 0;
-               for (i = 0; i < pools; i++) {
-                       vnic = bnxt_alloc_vnic(bp);
-                       if (!vnic) {
-                               RTE_LOG(ERR, PMD,
-                                       "VNIC alloc failed\n");
-                               rc = -ENOMEM;
-                               goto err_out;
-                       }
-                       vnic->flags |= BNXT_VNIC_INFO_BCAST;
-                       STAILQ_INSERT_TAIL(&bp->ff_pool[i], vnic, next);
-                       bp->nr_vnics++;
+               vnic->flags |= BNXT_VNIC_INFO_BCAST;
+               STAILQ_INSERT_TAIL(&bp->ff_pool[i], vnic, next);
+               bp->nr_vnics++;
 
-                       for (j = 0; j < nb_q_per_grp; j++, ring_idx++) {
-                               rxq = bp->eth_dev->data->rx_queues[ring_idx];
-                               rxq->vnic = vnic;
-                       }
-                       if (i == 0)
-                               vnic->func_default = true;
-                       vnic->ff_pool_idx = i;
-                       vnic->start_grp_id = start_grp_id;
-                       vnic->end_grp_id = end_grp_id;
-
-                       filter = bnxt_alloc_filter(bp);
-                       if (!filter) {
-                               RTE_LOG(ERR, PMD,
-                                       "L2 filter alloc failed\n");
-                               rc = -ENOMEM;
-                               goto err_out;
+               for (j = 0; j < nb_q_per_grp; j++, ring_idx++) {
+                       rxq = bp->eth_dev->data->rx_queues[ring_idx];
+                       rxq->vnic = vnic;
+               }
+               if (i == 0) {
+                       if (dev_conf->rxmode.mq_mode & ETH_MQ_RX_VMDQ_DCB) {
+                               bp->eth_dev->data->promiscuous = 1;
+                               vnic->flags |= BNXT_VNIC_INFO_PROMISC;
                        }
-                       /*
-                        * TODO: Configure & associate CFA rule for
-                        * each VNIC for each VMDq with MACVLAN, MACVLAN+TC
-                        */
-                       STAILQ_INSERT_TAIL(&vnic->filter, filter, next);
-
-                       start_grp_id = end_grp_id + 1;
-                       end_grp_id += nb_q_per_grp;
+                       vnic->func_default = true;
                }
-               goto out;
-       }
+               vnic->ff_pool_idx = i;
+               vnic->start_grp_id = start_grp_id;
+               vnic->end_grp_id = end_grp_id;
+
+               if (i) {
+                       if (dev_conf->rxmode.mq_mode & ETH_MQ_RX_VMDQ_DCB ||
+                           !(dev_conf->rxmode.mq_mode & ETH_MQ_RX_RSS))
+                               vnic->rss_dflt_cr = true;
+                       goto skip_filter_allocation;
+               }
+               filter = bnxt_alloc_filter(bp);
+               if (!filter) {
+                       RTE_LOG(ERR, PMD, "L2 filter alloc failed\n");
+                       rc = -ENOMEM;
+                       goto err_out;
+               }
+               /*
+                * TODO: Configure & associate CFA rule for
+                * each VNIC for each VMDq with MACVLAN, MACVLAN+TC
+                */
+               STAILQ_INSERT_TAIL(&vnic->filter, filter, next);
 
-       /* Non-VMDq mode - RSS, DCB, RSS+DCB */
-       /* Init default VNIC for RSS or DCB only */
-       vnic = bnxt_alloc_vnic(bp);
-       if (!vnic) {
-               RTE_LOG(ERR, PMD, "VNIC alloc failed\n");
-               rc = -ENOMEM;
-               goto err_out;
-       }
-       vnic->flags |= BNXT_VNIC_INFO_BCAST;
-       /* Partition the rx queues for the single pool */
-       for (i = 0; i < bp->rx_cp_nr_rings; i++) {
-               rxq = bp->eth_dev->data->rx_queues[i];
-               rxq->vnic = vnic;
-       }
-       STAILQ_INSERT_TAIL(&bp->ff_pool[0], vnic, next);
-       bp->nr_vnics++;
-
-       vnic->func_default = true;
-       vnic->ff_pool_idx = 0;
-       vnic->start_grp_id = 0;
-       vnic->end_grp_id = bp->rx_cp_nr_rings;
-       filter = bnxt_alloc_filter(bp);
-       if (!filter) {
-               RTE_LOG(ERR, PMD, "L2 filter alloc failed\n");
-               rc = -ENOMEM;
-               goto err_out;
+skip_filter_allocation:
+               start_grp_id = end_grp_id;
+               end_grp_id += nb_q_per_grp;
        }
-       STAILQ_INSERT_TAIL(&vnic->filter, filter, next);
-
-       if (dev_conf->rxmode.mq_mode & ETH_MQ_RX_RSS_FLAG)
-               vnic->hash_type =
-                       HWRM_VNIC_RSS_CFG_INPUT_HASH_TYPE_IPV4 |
-                       HWRM_VNIC_RSS_CFG_INPUT_HASH_TYPE_IPV6;
 
 out:
+       if (dev_conf->rxmode.mq_mode & ETH_MQ_RX_RSS_FLAG) {
+               struct rte_eth_rss_conf *rss = &dev_conf->rx_adv_conf.rss_conf;
+               uint16_t hash_type = 0;
+
+               if (bp->flags & BNXT_FLAG_UPDATE_HASH) {
+                       rss = &bp->rss_conf;
+                       bp->flags &= ~BNXT_FLAG_UPDATE_HASH;
+               }
+
+               if (rss->rss_hf & ETH_RSS_IPV4)
+                       hash_type |= HWRM_VNIC_RSS_CFG_INPUT_HASH_TYPE_IPV4;
+               if (rss->rss_hf & ETH_RSS_NONFRAG_IPV4_TCP)
+                       hash_type |= HWRM_VNIC_RSS_CFG_INPUT_HASH_TYPE_TCP_IPV4;
+               if (rss->rss_hf & ETH_RSS_NONFRAG_IPV4_UDP)
+                       hash_type |= HWRM_VNIC_RSS_CFG_INPUT_HASH_TYPE_UDP_IPV4;
+               if (rss->rss_hf & ETH_RSS_IPV6)
+                       hash_type |= HWRM_VNIC_RSS_CFG_INPUT_HASH_TYPE_IPV6;
+               if (rss->rss_hf & ETH_RSS_NONFRAG_IPV6_TCP)
+                       hash_type |= HWRM_VNIC_RSS_CFG_INPUT_HASH_TYPE_TCP_IPV6;
+               if (rss->rss_hf & ETH_RSS_NONFRAG_IPV6_UDP)
+                       hash_type |= HWRM_VNIC_RSS_CFG_INPUT_HASH_TYPE_UDP_IPV6;
+
+               for (i = 0; i < bp->nr_vnics; i++) {
+                       STAILQ_FOREACH(vnic, &bp->ff_pool[i], next) {
+                       vnic->hash_type = hash_type;
+
+                       /*
+                        * Use the supplied key if the key length is
+                        * acceptable and the rss_key is not NULL
+                        */
+                       if (rss->rss_key &&
+                           rss->rss_key_len <= HW_HASH_KEY_SIZE)
+                               memcpy(vnic->rss_hash_key,
+                                      rss->rss_key, rss->rss_key_len);
+                       }
+               }
+       }
+
        return rc;
 
 err_out:
@@ -349,3 +360,41 @@ int bnxt_rx_queue_setup_op(struct rte_eth_dev *eth_dev,
 out:
        return rc;
 }
+
+int
+bnxt_rx_queue_intr_enable_op(struct rte_eth_dev *eth_dev, uint16_t queue_id)
+{
+       struct bnxt_rx_queue *rxq;
+       struct bnxt_cp_ring_info *cpr;
+       int rc = 0;
+
+       if (eth_dev->data->rx_queues) {
+               rxq = eth_dev->data->rx_queues[queue_id];
+               if (!rxq) {
+                       rc = -EINVAL;
+                       return rc;
+               }
+               cpr = rxq->cp_ring;
+               B_CP_DB_ARM(cpr);
+       }
+       return rc;
+}
+
+int
+bnxt_rx_queue_intr_disable_op(struct rte_eth_dev *eth_dev, uint16_t queue_id)
+{
+       struct bnxt_rx_queue *rxq;
+       struct bnxt_cp_ring_info *cpr;
+       int rc = 0;
+
+       if (eth_dev->data->rx_queues) {
+               rxq = eth_dev->data->rx_queues[queue_id];
+               if (!rxq) {
+                       rc = -EINVAL;
+                       return rc;
+               }
+               cpr = rxq->cp_ring;
+               B_CP_DB_DISARM(cpr);
+       }
+       return rc;
+}