New upstream version 18.11.2
[deb_dpdk.git] / drivers / net / mlx5 / mlx5.c
index a277b57..d91d55b 100644 (file)
 /* Device parameter to enable RX completion queue compression. */
 #define MLX5_RXQ_CQE_COMP_EN "rxq_cqe_comp_en"
 
+/* Device parameter to enable RX completion entry padding to 128B. */
+#define MLX5_RXQ_CQE_PAD_EN "rxq_cqe_pad_en"
+
+/* Device parameter to enable padding Rx packet to cacheline size. */
+#define MLX5_RXQ_PKT_PAD_EN "rxq_pkt_pad_en"
+
 /* Device parameter to enable Multi-Packet Rx queue. */
 #define MLX5_RX_MPRQ_EN "mprq_en"
 
  */
 #define MLX5_TXQS_MIN_INLINE "txqs_min_inline"
 
+/*
+ * Device parameter to configure the number of TX queues threshold for
+ * enabling vectorized Tx.
+ */
+#define MLX5_TXQS_MAX_VEC "txqs_max_vec"
+
 /* Device parameter to enable multi-packet send WQEs. */
 #define MLX5_TXQ_MPW_EN "txq_mpw_en"
 
@@ -145,9 +157,10 @@ mlx5_prepare_shared_data(void)
                if (rte_eal_process_type() == RTE_PROC_PRIMARY) {
                        LIST_INIT(&mlx5_shared_data->mem_event_cb_list);
                        rte_rwlock_init(&mlx5_shared_data->mem_event_rwlock);
+                       rte_mem_event_callback_register("MLX5_MEM_EVENT_CB",
+                                                       mlx5_mr_mem_event_cb,
+                                                       NULL);
                }
-               rte_mem_event_callback_register("MLX5_MEM_EVENT_CB",
-                                               mlx5_mr_mem_event_cb, NULL);
        }
        rte_spinlock_unlock(&mlx5_shared_data_lock);
 }
@@ -188,7 +201,7 @@ mlx5_getenv_int(const char *name)
 static void *
 mlx5_alloc_verbs_buf(size_t size, void *data)
 {
-       struct priv *priv = data;
+       struct mlx5_priv *priv = data;
        void *ret;
        size_t alignment = sysconf(_SC_PAGESIZE);
        unsigned int socket = SOCKET_ID_ANY;
@@ -236,7 +249,7 @@ mlx5_free_verbs_buf(void *ptr, void *data __rte_unused)
 static void
 mlx5_dev_close(struct rte_eth_dev *dev)
 {
-       struct priv *priv = dev->data->dev_private;
+       struct mlx5_priv *priv = dev->data->dev_private;
        unsigned int i;
        int ret;
 
@@ -323,7 +336,7 @@ mlx5_dev_close(struct rte_eth_dev *dev)
 
                i = RTE_MIN(mlx5_dev_to_port_id(dev->device, port_id, i), i);
                while (i--) {
-                       struct priv *opriv =
+                       struct mlx5_priv *opriv =
                                rte_eth_devices[port_id[i]].data->dev_private;
 
                        if (!opriv ||
@@ -337,11 +350,6 @@ mlx5_dev_close(struct rte_eth_dev *dev)
        }
        memset(priv, 0, sizeof(*priv));
        priv->domain_id = RTE_ETH_DEV_SWITCH_DOMAIN_ID_INVALID;
-       /*
-        * flag to rte_eth_dev_close() that it should release the port resources
-        * (calling rte_eth_dev_release_port()) in addition to closing it.
-        */
-       dev->data->dev_flags |= RTE_ETH_DEV_CLOSE_REMOVE;
        /*
         * Reset mac_addrs to NULL such that it is not freed as part of
         * rte_eth_dev_release_port(). mac_addrs is part of dev_private so
@@ -390,6 +398,7 @@ const struct eth_dev_ops mlx5_dev_ops = {
        .filter_ctrl = mlx5_dev_filter_ctrl,
        .rx_descriptor_status = mlx5_rx_descriptor_status,
        .tx_descriptor_status = mlx5_tx_descriptor_status,
+       .rx_queue_count = mlx5_rx_queue_count,
        .rx_queue_intr_enable = mlx5_rx_intr_enable,
        .rx_queue_intr_disable = mlx5_rx_intr_disable,
        .is_removed = mlx5_is_removed,
@@ -479,6 +488,10 @@ mlx5_args_check(const char *key, const char *val, void *opaque)
        }
        if (strcmp(MLX5_RXQ_CQE_COMP_EN, key) == 0) {
                config->cqe_comp = !!tmp;
+       } else if (strcmp(MLX5_RXQ_CQE_PAD_EN, key) == 0) {
+               config->cqe_pad = !!tmp;
+       } else if (strcmp(MLX5_RXQ_PKT_PAD_EN, key) == 0) {
+               config->hw_padding = !!tmp;
        } else if (strcmp(MLX5_RX_MPRQ_EN, key) == 0) {
                config->mprq.enabled = !!tmp;
        } else if (strcmp(MLX5_RX_MPRQ_LOG_STRIDE_NUM, key) == 0) {
@@ -491,6 +504,8 @@ mlx5_args_check(const char *key, const char *val, void *opaque)
                config->txq_inline = tmp;
        } else if (strcmp(MLX5_TXQS_MIN_INLINE, key) == 0) {
                config->txqs_inline = tmp;
+       } else if (strcmp(MLX5_TXQS_MAX_VEC, key) == 0) {
+               config->txqs_vec = tmp;
        } else if (strcmp(MLX5_TXQ_MPW_EN, key) == 0) {
                config->mps = !!tmp;
        } else if (strcmp(MLX5_TXQ_MPW_HDR_DSEG_EN, key) == 0) {
@@ -531,12 +546,15 @@ mlx5_args(struct mlx5_dev_config *config, struct rte_devargs *devargs)
 {
        const char **params = (const char *[]){
                MLX5_RXQ_CQE_COMP_EN,
+               MLX5_RXQ_CQE_PAD_EN,
+               MLX5_RXQ_PKT_PAD_EN,
                MLX5_RX_MPRQ_EN,
                MLX5_RX_MPRQ_LOG_STRIDE_NUM,
                MLX5_RX_MPRQ_MAX_MEMCPY_LEN,
                MLX5_RXQS_MIN_MPRQ,
                MLX5_TXQ_INLINE,
                MLX5_TXQS_MIN_INLINE,
+               MLX5_TXQS_MAX_VEC,
                MLX5_TXQ_MPW_EN,
                MLX5_TXQ_MPW_HDR_DSEG_EN,
                MLX5_TXQ_MAX_INLINE_LEN,
@@ -613,7 +631,7 @@ find_lower_va_bound(const struct rte_memseg_list *msl,
 static int
 mlx5_uar_init_primary(struct rte_eth_dev *dev)
 {
-       struct priv *priv = dev->data->dev_private;
+       struct mlx5_priv *priv = dev->data->dev_private;
        void *addr = (void *)0;
 
        if (uar_base) { /* UAR address space mapped. */
@@ -659,7 +677,7 @@ mlx5_uar_init_primary(struct rte_eth_dev *dev)
 static int
 mlx5_uar_init_secondary(struct rte_eth_dev *dev)
 {
-       struct priv *priv = dev->data->dev_private;
+       struct mlx5_priv *priv = dev->data->dev_private;
        void *addr;
 
        assert(priv->uar_base);
@@ -698,8 +716,8 @@ mlx5_uar_init_secondary(struct rte_eth_dev *dev)
  *   Backing DPDK device.
  * @param ibv_dev
  *   Verbs device.
- * @param vf
- *   If nonzero, enable VF-specific features.
+ * @param config
+ *   Device configuration parameters.
  * @param[in] switch_info
  *   Switch properties of Ethernet device.
  *
@@ -713,7 +731,7 @@ mlx5_uar_init_secondary(struct rte_eth_dev *dev)
 static struct rte_eth_dev *
 mlx5_dev_spawn(struct rte_device *dpdk_dev,
               struct ibv_device *ibv_dev,
-              int vf,
+              struct mlx5_dev_config config,
               const struct mlx5_switch_info *switch_info)
 {
        struct ibv_context *ctx;
@@ -721,28 +739,13 @@ mlx5_dev_spawn(struct rte_device *dpdk_dev,
        struct ibv_port_attr port_attr;
        struct ibv_pd *pd = NULL;
        struct mlx5dv_context dv_attr = { .comp_mask = 0 };
-       struct mlx5_dev_config config = {
-               .vf = !!vf,
-               .mps = MLX5_ARG_UNSET,
-               .tx_vec_en = 1,
-               .rx_vec_en = 1,
-               .mpw_hdr_dseg = 0,
-               .txq_inline = MLX5_ARG_UNSET,
-               .txqs_inline = MLX5_ARG_UNSET,
-               .inline_max_packet_sz = MLX5_ARG_UNSET,
-               .vf_nl_en = 1,
-               .mprq = {
-                       .enabled = 0,
-                       .stride_num_n = MLX5_MPRQ_STRIDE_NUM_N,
-                       .max_memcpy_len = MLX5_MPRQ_MEMCPY_DEFAULT_LEN,
-                       .min_rxqs_num = MLX5_MPRQ_MIN_RXQS,
-               },
-       };
        struct rte_eth_dev *eth_dev = NULL;
-       struct priv *priv = NULL;
+       struct mlx5_priv *priv = NULL;
        int err = 0;
+       unsigned int hw_padding = 0;
        unsigned int mps;
        unsigned int cqe_comp;
+       unsigned int cqe_pad = 0;
        unsigned int tunnel_en = 0;
        unsigned int mpls_en = 0;
        unsigned int swp = 0;
@@ -863,6 +866,11 @@ mlx5_dev_spawn(struct rte_device *dpdk_dev,
        else
                cqe_comp = 1;
        config.cqe_comp = cqe_comp;
+#ifdef HAVE_IBV_MLX5_MOD_CQE_128B_PAD
+       /* Whether device supports 128B Rx CQE padding. */
+       cqe_pad = RTE_CACHE_LINE_SIZE == 128 &&
+                 (dv_attr.flags & MLX5DV_CONTEXT_FLAGS_CQE_128B_PAD);
+#endif
 #ifdef HAVE_IBV_DEVICE_TUNNEL_SUPPORT
        if (dv_attr.comp_mask & MLX5DV_CONTEXT_MASK_TUNNEL_OFFLOADS) {
                tunnel_en = ((dv_attr.tunnel_offloads_caps &
@@ -994,7 +1002,7 @@ mlx5_dev_spawn(struct rte_device *dpdk_dev,
 
                i = RTE_MIN(mlx5_dev_to_port_id(dpdk_dev, port_id, i), i);
                while (i--) {
-                       const struct priv *opriv =
+                       const struct mlx5_priv *opriv =
                                rte_eth_devices[port_id[i]].data->dev_private;
 
                        if (!opriv ||
@@ -1053,11 +1061,18 @@ mlx5_dev_spawn(struct rte_device *dpdk_dev,
                                 IBV_RAW_PACKET_CAP_SCATTER_FCS);
        DRV_LOG(DEBUG, "FCS stripping configuration is %ssupported",
                (config.hw_fcs_strip ? "" : "not "));
-#ifdef HAVE_IBV_WQ_FLAG_RX_END_PADDING
-       config.hw_padding = !!attr.rx_pad_end_addr_align;
+#if defined(HAVE_IBV_WQ_FLAG_RX_END_PADDING)
+       hw_padding = !!attr.rx_pad_end_addr_align;
+#elif defined(HAVE_IBV_WQ_FLAGS_PCI_WRITE_END_PADDING)
+       hw_padding = !!(attr.device_cap_flags_ex &
+                       IBV_DEVICE_PCI_WRITE_END_PADDING);
 #endif
-       DRV_LOG(DEBUG, "hardware Rx end alignment padding is %ssupported",
-               (config.hw_padding ? "" : "not "));
+       if (config.hw_padding && !hw_padding) {
+               DRV_LOG(DEBUG, "Rx end alignment padding isn't supported");
+               config.hw_padding = 0;
+       } else if (config.hw_padding) {
+               DRV_LOG(DEBUG, "Rx end alignment padding is enabled");
+       }
        config.tso = (attr.tso_caps.max_tso > 0 &&
                      (attr.tso_caps.supported_qpts &
                       (1 << IBV_QPT_RAW_PACKET)));
@@ -1079,6 +1094,12 @@ mlx5_dev_spawn(struct rte_device *dpdk_dev,
                DRV_LOG(WARNING, "Rx CQE compression isn't supported");
                config.cqe_comp = 0;
        }
+       if (config.cqe_pad && !cqe_pad) {
+               DRV_LOG(WARNING, "Rx CQE padding isn't supported");
+               config.cqe_pad = 0;
+       } else if (config.cqe_pad) {
+               DRV_LOG(INFO, "Rx CQE padding is enabled");
+       }
        if (config.mprq.enabled && mprq) {
                if (config.mprq.stride_num_n > mprq_max_stride_num_n ||
                    config.mprq.stride_num_n < mprq_min_stride_num_n) {
@@ -1103,6 +1124,8 @@ mlx5_dev_spawn(struct rte_device *dpdk_dev,
                err = ENOMEM;
                goto error;
        }
+       /* Flag to call rte_eth_dev_release_port() in rte_eth_dev_close(). */
+       eth_dev->data->dev_flags |= RTE_ETH_DEV_CLOSE_REMOVE;
        if (priv->representor) {
                eth_dev->data->dev_flags |= RTE_ETH_DEV_REPRESENTOR;
                eth_dev->data->representor_id = priv->representor_id;
@@ -1157,7 +1180,7 @@ mlx5_dev_spawn(struct rte_device *dpdk_dev,
        eth_dev->dev_ops = &mlx5_dev_ops;
        /* Register MAC address. */
        claim_zero(mlx5_mac_addr_add(eth_dev, &mac, 0, 0));
-       if (vf && config.vf_nl_en)
+       if (config.vf && config.vf_nl_en)
                mlx5_nl_mac_addr_sync(eth_dev);
        priv->tcf_context = mlx5_flow_tcf_context_create();
        if (!priv->tcf_context) {
@@ -1211,8 +1234,10 @@ mlx5_dev_spawn(struct rte_device *dpdk_dev,
        priv->config = config;
        /* Supported Verbs flow priority number detection. */
        err = mlx5_flow_discover_priorities(eth_dev);
-       if (err < 0)
+       if (err < 0) {
+               err = -err;
                goto error;
+       }
        priv->config.flow_prio = err;
        /*
         * Once the device is added to the list of memory event
@@ -1326,7 +1351,7 @@ mlx5_pci_probe(struct rte_pci_driver *pci_drv __rte_unused,
 {
        struct ibv_device **ibv_list;
        unsigned int n = 0;
-       int vf;
+       struct mlx5_dev_config dev_config;
        int ret;
 
        assert(pci_drv == &mlx5_driver);
@@ -1424,21 +1449,47 @@ mlx5_pci_probe(struct rte_pci_driver *pci_drv __rte_unused,
         */
        if (n)
                qsort(list, n, sizeof(*list), mlx5_dev_spawn_data_cmp);
+       /* Default configuration. */
+       dev_config = (struct mlx5_dev_config){
+               .hw_padding = 0,
+               .mps = MLX5_ARG_UNSET,
+               .tx_vec_en = 1,
+               .rx_vec_en = 1,
+               .txq_inline = MLX5_ARG_UNSET,
+               .txqs_inline = MLX5_ARG_UNSET,
+               .txqs_vec = MLX5_ARG_UNSET,
+               .inline_max_packet_sz = MLX5_ARG_UNSET,
+               .vf_nl_en = 1,
+               .mprq = {
+                       .enabled = 0, /* Disabled by default. */
+                       .stride_num_n = MLX5_MPRQ_STRIDE_NUM_N,
+                       .max_memcpy_len = MLX5_MPRQ_MEMCPY_DEFAULT_LEN,
+                       .min_rxqs_num = MLX5_MPRQ_MIN_RXQS,
+               },
+       };
+       /* Device speicific configuration. */
        switch (pci_dev->id.device_id) {
+       case PCI_DEVICE_ID_MELLANOX_CONNECTX5BF:
+               dev_config.txqs_vec = MLX5_VPMD_MAX_TXQS_BLUEFIELD;
+               break;
        case PCI_DEVICE_ID_MELLANOX_CONNECTX4VF:
        case PCI_DEVICE_ID_MELLANOX_CONNECTX4LXVF:
        case PCI_DEVICE_ID_MELLANOX_CONNECTX5VF:
        case PCI_DEVICE_ID_MELLANOX_CONNECTX5EXVF:
-               vf = 1;
+               dev_config.vf = 1;
                break;
        default:
-               vf = 0;
+               break;
        }
+       /* Set architecture-dependent default value if unset. */
+       if (dev_config.txqs_vec == MLX5_ARG_UNSET)
+               dev_config.txqs_vec = MLX5_VPMD_MAX_TXQS;
        for (i = 0; i != n; ++i) {
                uint32_t restore;
 
-               list[i].eth_dev = mlx5_dev_spawn
-                       (&pci_dev->device, list[i].ibv_dev, vf, &list[i].info);
+               list[i].eth_dev = mlx5_dev_spawn(&pci_dev->device,
+                                                list[i].ibv_dev, dev_config,
+                                                &list[i].info);
                if (!list[i].eth_dev) {
                        if (rte_errno != EBUSY && rte_errno != EEXIST)
                                break;