New upstream version 18.02
[deb_dpdk.git] / drivers / net / enic / enic_ethdev.c
index 6bea940..d84714e 100644 (file)
@@ -1,35 +1,6 @@
-/*
- * Copyright 2008-2014 Cisco Systems, Inc.  All rights reserved.
+/* SPDX-License-Identifier: BSD-3-Clause
+ * Copyright 2008-2017 Cisco Systems, Inc.  All rights reserved.
  * Copyright 2007 Nuova Systems, Inc.  All rights reserved.
- *
- * Copyright (c) 2014, Cisco Systems, Inc.
- * All rights reserved.
- *
- * Redistribution and use in source and binary forms, with or without
- * modification, are permitted provided that the following conditions
- * are met:
- *
- * 1. Redistributions of source code must retain the above copyright
- * notice, this list of conditions and the following disclaimer.
- *
- * 2. Redistributions in binary form must reproduce the above copyright
- * notice, this list of conditions and the following disclaimer in
- * the documentation and/or other materials provided with the
- * distribution.
- *
- * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
- * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
- * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS
- * FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE
- * COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT,
- * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING,
- * BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
- * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER
- * CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
- * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN
- * ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
- * POSSIBILITY OF SUCH DAMAGE.
- *
  */
 
 #include <stdio.h>
@@ -37,7 +8,9 @@
 
 #include <rte_dev.h>
 #include <rte_pci.h>
-#include <rte_ethdev.h>
+#include <rte_bus_pci.h>
+#include <rte_ethdev_driver.h>
+#include <rte_ethdev_pci.h>
 #include <rte_string_fns.h>
 
 #include "vnic_intr.h"
 #include "vnic_enet.h"
 #include "enic.h"
 
-#ifdef RTE_LIBRTE_ENIC_DEBUG
-#define ENICPMD_FUNC_TRACE() \
-       RTE_LOG(DEBUG, PMD, "ENICPMD trace: %s\n", __func__)
-#else
-#define ENICPMD_FUNC_TRACE() (void)0
-#endif
+int enicpmd_logtype_init;
+int enicpmd_logtype_flow;
+
+#define PMD_INIT_LOG(level, fmt, args...) \
+       rte_log(RTE_LOG_ ## level, enicpmd_logtype_init, \
+               "%s" fmt "\n", __func__, ##args)
+
+#define ENICPMD_FUNC_TRACE() PMD_INIT_LOG(DEBUG, " >>")
 
 /*
  * The set of PCI devices this driver supports
  */
+#define CISCO_PCI_VENDOR_ID 0x1137
 static const struct rte_pci_id pci_id_enic_map[] = {
-#define RTE_PCI_DEV_ID_DECL_ENIC(vend, dev) {RTE_PCI_DEVICE(vend, dev)},
-#ifndef PCI_VENDOR_ID_CISCO
-#define PCI_VENDOR_ID_CISCO    0x1137
-#endif
-#include "rte_pci_dev_ids.h"
-RTE_PCI_DEV_ID_DECL_ENIC(PCI_VENDOR_ID_CISCO, PCI_DEVICE_ID_CISCO_VIC_ENET)
-RTE_PCI_DEV_ID_DECL_ENIC(PCI_VENDOR_ID_CISCO, PCI_DEVICE_ID_CISCO_VIC_ENET_VF)
-{.vendor_id = 0, /* Sentinal */},
+       { RTE_PCI_DEVICE(CISCO_PCI_VENDOR_ID, PCI_DEVICE_ID_CISCO_VIC_ENET) },
+       { RTE_PCI_DEVICE(CISCO_PCI_VENDOR_ID, PCI_DEVICE_ID_CISCO_VIC_ENET_VF) },
+       {.vendor_id = 0, /* sentinel */},
 };
 
+RTE_INIT(enicpmd_init_log);
+static void
+enicpmd_init_log(void)
+{
+       enicpmd_logtype_init = rte_log_register("pmd.net.enic.init");
+       if (enicpmd_logtype_init >= 0)
+               rte_log_set_level(enicpmd_logtype_init, RTE_LOG_NOTICE);
+       enicpmd_logtype_flow = rte_log_register("pmd.net.enic.flow");
+       if (enicpmd_logtype_flow >= 0)
+               rte_log_set_level(enicpmd_logtype_flow, RTE_LOG_NOTICE);
+}
+
 static int
 enicpmd_fdir_ctrl_func(struct rte_eth_dev *eth_dev,
                        enum rte_filter_op filter_op, void *arg)
@@ -99,10 +82,12 @@ enicpmd_fdir_ctrl_func(struct rte_eth_dev *eth_dev,
                break;
 
        case RTE_ETH_FILTER_FLUSH:
-       case RTE_ETH_FILTER_INFO:
                dev_warning(enic, "unsupported operation %u", filter_op);
                ret = -ENOTSUP;
                break;
+       case RTE_ETH_FILTER_INFO:
+               enic_fdir_info_get(enic, (struct rte_eth_fdir_info *)arg);
+               break;
        default:
                dev_err(enic, "unknown operation %u", filter_op);
                ret = -EINVAL;
@@ -117,13 +102,25 @@ enicpmd_dev_filter_ctrl(struct rte_eth_dev *dev,
                     enum rte_filter_op filter_op,
                     void *arg)
 {
-       int ret = -EINVAL;
+       int ret = 0;
 
-       if (RTE_ETH_FILTER_FDIR == filter_type)
+       ENICPMD_FUNC_TRACE();
+
+       switch (filter_type) {
+       case RTE_ETH_FILTER_GENERIC:
+               if (filter_op != RTE_ETH_FILTER_GET)
+                       return -EINVAL;
+               *(const void **)arg = &enic_flow_ops;
+               break;
+       case RTE_ETH_FILTER_FDIR:
                ret = enicpmd_fdir_ctrl_func(dev, filter_op, arg);
-       else
+               break;
+       default:
                dev_warning(enic, "Filter type (%d) not supported",
                        filter_type);
+               ret = -EINVAL;
+               break;
+       }
 
        return ret;
 }
@@ -131,6 +128,10 @@ enicpmd_dev_filter_ctrl(struct rte_eth_dev *dev,
 static void enicpmd_dev_tx_queue_release(void *txq)
 {
        ENICPMD_FUNC_TRACE();
+
+       if (rte_eal_process_type() != RTE_PROC_PRIMARY)
+               return;
+
        enic_free_wq(txq);
 }
 
@@ -146,9 +147,21 @@ static int enicpmd_dev_setup_intr(struct enic *enic)
                if (!enic->cq[index].ctrl)
                        break;
        }
-
        if (enic->cq_count != index)
                return 0;
+       for (index = 0; index < enic->wq_count; index++) {
+               if (!enic->wq[index].ctrl)
+                       break;
+       }
+       if (enic->wq_count != index)
+               return 0;
+       /* check start of packet (SOP) RQs only in case scatter is disabled. */
+       for (index = 0; index < enic->rq_count; index++) {
+               if (!enic->rq[enic_rte_rq_idx_to_sop_idx(index)].ctrl)
+                       break;
+       }
+       if (enic->rq_count != index)
+               return 0;
 
        ret = enic_alloc_intr_resources(enic);
        if (ret) {
@@ -173,14 +186,11 @@ static int enicpmd_dev_tx_queue_setup(struct rte_eth_dev *eth_dev,
        int ret;
        struct enic *enic = pmd_priv(eth_dev);
 
-       ENICPMD_FUNC_TRACE();
-       if (queue_idx >= ENIC_WQ_MAX) {
-               dev_err(enic,
-                       "Max number of TX queues exceeded.  Max is %d\n",
-                       ENIC_WQ_MAX);
-               return -EINVAL;
-       }
+       if (rte_eal_process_type() != RTE_PROC_PRIMARY)
+               return -E_RTE_SECONDARY;
 
+       ENICPMD_FUNC_TRACE();
+       RTE_ASSERT(queue_idx < enic->conf_wq_count);
        eth_dev->data->tx_queues[queue_idx] = (void *)&enic->wq[queue_idx];
 
        ret = enic_alloc_wq(enic, queue_idx, socket_id, nb_desc);
@@ -200,7 +210,6 @@ static int enicpmd_dev_tx_queue_start(struct rte_eth_dev *eth_dev,
        ENICPMD_FUNC_TRACE();
 
        enic_start_wq(enic, queue_idx);
-       eth_dev->data->tx_queue_state[queue_idx] = RTE_ETH_QUEUE_STATE_STARTED;
 
        return 0;
 }
@@ -216,8 +225,6 @@ static int enicpmd_dev_tx_queue_stop(struct rte_eth_dev *eth_dev,
        ret = enic_stop_wq(enic, queue_idx);
        if (ret)
                dev_err(enic, "error in stopping wq %d\n", queue_idx);
-       else
-               eth_dev->data->tx_queue_state[queue_idx] = RTE_ETH_QUEUE_STATE_STOPPED;
 
        return ret;
 }
@@ -230,7 +237,6 @@ static int enicpmd_dev_rx_queue_start(struct rte_eth_dev *eth_dev,
        ENICPMD_FUNC_TRACE();
 
        enic_start_rq(enic, queue_idx);
-       eth_dev->data->rx_queue_state[queue_idx] = RTE_ETH_QUEUE_STATE_STARTED;
 
        return 0;
 }
@@ -246,8 +252,6 @@ static int enicpmd_dev_rx_queue_stop(struct rte_eth_dev *eth_dev,
        ret = enic_stop_rq(enic, queue_idx);
        if (ret)
                dev_err(enic, "error in stopping rq %d\n", queue_idx);
-       else
-               eth_dev->data->rx_queue_state[queue_idx] = RTE_ETH_QUEUE_STATE_STOPPED;
 
        return ret;
 }
@@ -255,9 +259,37 @@ static int enicpmd_dev_rx_queue_stop(struct rte_eth_dev *eth_dev,
 static void enicpmd_dev_rx_queue_release(void *rxq)
 {
        ENICPMD_FUNC_TRACE();
+
+       if (rte_eal_process_type() != RTE_PROC_PRIMARY)
+               return;
+
        enic_free_rq(rxq);
 }
 
+static uint32_t enicpmd_dev_rx_queue_count(struct rte_eth_dev *dev,
+                                          uint16_t rx_queue_id)
+{
+       struct enic *enic = pmd_priv(dev);
+       uint32_t queue_count = 0;
+       struct vnic_cq *cq;
+       uint32_t cq_tail;
+       uint16_t cq_idx;
+       int rq_num;
+
+       rq_num = enic_rte_rq_idx_to_sop_idx(rx_queue_id);
+       cq = &enic->cq[enic_cq_rq(enic, rq_num)];
+       cq_idx = cq->to_clean;
+
+       cq_tail = ioread32(&cq->ctrl->cq_tail);
+
+       if (cq_tail < cq_idx)
+               cq_tail += cq->ring.desc_count;
+
+       queue_count = cq_tail - cq_idx;
+
+       return queue_count;
+}
+
 static int enicpmd_dev_rx_queue_setup(struct rte_eth_dev *eth_dev,
        uint16_t queue_idx,
        uint16_t nb_desc,
@@ -269,25 +301,20 @@ static int enicpmd_dev_rx_queue_setup(struct rte_eth_dev *eth_dev,
        struct enic *enic = pmd_priv(eth_dev);
 
        ENICPMD_FUNC_TRACE();
-       if (queue_idx >= ENIC_RQ_MAX) {
-               dev_err(enic,
-                       "Max number of RX queues exceeded.  Max is %d\n",
-                       ENIC_RQ_MAX);
-               return -EINVAL;
-       }
 
-       eth_dev->data->rx_queues[queue_idx] = (void *)&enic->rq[queue_idx];
+       if (rte_eal_process_type() != RTE_PROC_PRIMARY)
+               return -E_RTE_SECONDARY;
+       RTE_ASSERT(enic_rte_rq_idx_to_sop_idx(queue_idx) < enic->conf_rq_count);
+       eth_dev->data->rx_queues[queue_idx] =
+               (void *)&enic->rq[enic_rte_rq_idx_to_sop_idx(queue_idx)];
 
-       ret = enic_alloc_rq(enic, queue_idx, socket_id, mp, nb_desc);
+       ret = enic_alloc_rq(enic, queue_idx, socket_id, mp, nb_desc,
+                           rx_conf->rx_free_thresh);
        if (ret) {
                dev_err(enic, "error in allocating rq\n");
                return ret;
        }
 
-       enic->rq[queue_idx].rx_free_thresh = rx_conf->rx_free_thresh;
-       dev_debug(enic, "Set queue_id:%u free thresh:%u\n", queue_idx,
-                       enic->rq[queue_idx].rx_free_thresh);
-
        return enicpmd_dev_setup_intr(enic);
 }
 
@@ -305,14 +332,15 @@ static int enicpmd_vlan_filter_set(struct rte_eth_dev *eth_dev,
        return err;
 }
 
-static void enicpmd_vlan_offload_set(struct rte_eth_dev *eth_dev, int mask)
+static int enicpmd_vlan_offload_set(struct rte_eth_dev *eth_dev, int mask)
 {
        struct enic *enic = pmd_priv(eth_dev);
 
        ENICPMD_FUNC_TRACE();
 
        if (mask & ETH_VLAN_STRIP_MASK) {
-               if (eth_dev->data->dev_conf.rxmode.hw_vlan_strip)
+               if (eth_dev->data->dev_conf.rxmode.offloads &
+                   DEV_RX_OFFLOAD_VLAN_STRIP)
                        enic->ig_vlan_strip_en = 1;
                else
                        enic->ig_vlan_strip_en = 0;
@@ -329,6 +357,8 @@ static void enicpmd_vlan_offload_set(struct rte_eth_dev *eth_dev, int mask)
                dev_warning(enic,
                        "Configuration of extended VLAN is not supported\n");
        }
+
+       return 0;
 }
 
 static int enicpmd_dev_configure(struct rte_eth_dev *eth_dev)
@@ -336,6 +366,9 @@ static int enicpmd_dev_configure(struct rte_eth_dev *eth_dev)
        int ret;
        struct enic *enic = pmd_priv(eth_dev);
 
+       if (rte_eal_process_type() != RTE_PROC_PRIMARY)
+               return -E_RTE_SECONDARY;
+
        ENICPMD_FUNC_TRACE();
        ret = enic_set_vnic_res(enic);
        if (ret) {
@@ -343,15 +376,11 @@ static int enicpmd_dev_configure(struct rte_eth_dev *eth_dev)
                return ret;
        }
 
-       if (eth_dev->data->dev_conf.rxmode.split_hdr_size &&
-               eth_dev->data->dev_conf.rxmode.header_split) {
-               /* Enable header-data-split */
-               enic_set_hdr_split_size(enic,
-                       eth_dev->data->dev_conf.rxmode.split_hdr_size);
-       }
+       enic->hw_ip_checksum = !!(eth_dev->data->dev_conf.rxmode.offloads &
+                                 DEV_RX_OFFLOAD_CHECKSUM);
+       ret = enicpmd_vlan_offload_set(eth_dev, ETH_VLAN_STRIP_MASK);
 
-       enic->hw_ip_checksum = eth_dev->data->dev_conf.rxmode.hw_ip_checksum;
-       return 0;
+       return ret;
 }
 
 /* Start the device.
@@ -361,6 +390,9 @@ static int enicpmd_dev_start(struct rte_eth_dev *eth_dev)
 {
        struct enic *enic = pmd_priv(eth_dev);
 
+       if (rte_eal_process_type() != RTE_PROC_PRIMARY)
+               return -E_RTE_SECONDARY;
+
        ENICPMD_FUNC_TRACE();
        return enic_enable(enic);
 }
@@ -373,6 +405,9 @@ static void enicpmd_dev_stop(struct rte_eth_dev *eth_dev)
        struct rte_eth_link link;
        struct enic *enic = pmd_priv(eth_dev);
 
+       if (rte_eal_process_type() != RTE_PROC_PRIMARY)
+               return;
+
        ENICPMD_FUNC_TRACE();
        enic_disable(enic);
        memset(&link, 0, sizeof(link));
@@ -396,26 +431,18 @@ static int enicpmd_dev_link_update(struct rte_eth_dev *eth_dev,
        __rte_unused int wait_to_complete)
 {
        struct enic *enic = pmd_priv(eth_dev);
-       int ret;
-       int link_status = 0;
 
        ENICPMD_FUNC_TRACE();
-       link_status = enic_get_link_status(enic);
-       ret = (link_status == enic->link_status);
-       enic->link_status = link_status;
-       eth_dev->data->dev_link.link_status = link_status;
-       eth_dev->data->dev_link.link_duplex = ETH_LINK_FULL_DUPLEX;
-       eth_dev->data->dev_link.link_speed = vnic_dev_port_speed(enic->vdev);
-       return ret;
+       return enic_link_update(enic);
 }
 
-static void enicpmd_dev_stats_get(struct rte_eth_dev *eth_dev,
+static int enicpmd_dev_stats_get(struct rte_eth_dev *eth_dev,
        struct rte_eth_stats *stats)
 {
        struct enic *enic = pmd_priv(eth_dev);
 
        ENICPMD_FUNC_TRACE();
-       enic_dev_stats_get(enic, stats);
+       return enic_dev_stats_get(enic, stats);
 }
 
 static void enicpmd_dev_stats_reset(struct rte_eth_dev *eth_dev)
@@ -432,11 +459,13 @@ static void enicpmd_dev_info_get(struct rte_eth_dev *eth_dev,
        struct enic *enic = pmd_priv(eth_dev);
 
        ENICPMD_FUNC_TRACE();
-       device_info->max_rx_queues = enic->rq_count;
-       device_info->max_tx_queues = enic->wq_count;
+       device_info->pci_dev = RTE_ETH_DEV_TO_PCI(eth_dev);
+       /* Scattered Rx uses two receive queues per rx queue exposed to dpdk */
+       device_info->max_rx_queues = enic->conf_rq_count / 2;
+       device_info->max_tx_queues = enic->conf_wq_count;
        device_info->min_rx_bufsize = ENIC_MIN_MTU;
-       device_info->max_rx_pktlen = enic->config.mtu;
-       device_info->max_mac_addrs = 1;
+       device_info->max_rx_pktlen = enic->max_mtu + ETHER_HDR_LEN + 4;
+       device_info->max_mac_addrs = ENIC_MAX_MAC_ADDR;
        device_info->rx_offload_capa =
                DEV_RX_OFFLOAD_VLAN_STRIP |
                DEV_RX_OFFLOAD_IPV4_CKSUM |
@@ -446,7 +475,8 @@ static void enicpmd_dev_info_get(struct rte_eth_dev *eth_dev,
                DEV_TX_OFFLOAD_VLAN_INSERT |
                DEV_TX_OFFLOAD_IPV4_CKSUM  |
                DEV_TX_OFFLOAD_UDP_CKSUM   |
-               DEV_TX_OFFLOAD_TCP_CKSUM;
+               DEV_TX_OFFLOAD_TCP_CKSUM   |
+               DEV_TX_OFFLOAD_TCP_TSO;
        device_info->default_rxconf = (struct rte_eth_rxconf) {
                .rx_free_thresh = ENIC_DEFAULT_RX_FREE_THRESH
        };
@@ -455,8 +485,14 @@ static void enicpmd_dev_info_get(struct rte_eth_dev *eth_dev,
 static const uint32_t *enicpmd_dev_supported_ptypes_get(struct rte_eth_dev *dev)
 {
        static const uint32_t ptypes[] = {
-               RTE_PTYPE_L3_IPV4,
-               RTE_PTYPE_L3_IPV6,
+               RTE_PTYPE_L2_ETHER,
+               RTE_PTYPE_L2_ETHER_VLAN,
+               RTE_PTYPE_L3_IPV4_EXT_UNKNOWN,
+               RTE_PTYPE_L3_IPV6_EXT_UNKNOWN,
+               RTE_PTYPE_L4_TCP,
+               RTE_PTYPE_L4_UDP,
+               RTE_PTYPE_L4_FRAG,
+               RTE_PTYPE_L4_NONFRAG,
                RTE_PTYPE_UNKNOWN
        };
 
@@ -469,7 +505,11 @@ static void enicpmd_dev_promiscuous_enable(struct rte_eth_dev *eth_dev)
 {
        struct enic *enic = pmd_priv(eth_dev);
 
+       if (rte_eal_process_type() != RTE_PROC_PRIMARY)
+               return;
+
        ENICPMD_FUNC_TRACE();
+
        enic->promisc = 1;
        enic_add_packet_filter(enic);
 }
@@ -478,6 +518,9 @@ static void enicpmd_dev_promiscuous_disable(struct rte_eth_dev *eth_dev)
 {
        struct enic *enic = pmd_priv(eth_dev);
 
+       if (rte_eal_process_type() != RTE_PROC_PRIMARY)
+               return;
+
        ENICPMD_FUNC_TRACE();
        enic->promisc = 0;
        enic_add_packet_filter(enic);
@@ -487,6 +530,9 @@ static void enicpmd_dev_allmulticast_enable(struct rte_eth_dev *eth_dev)
 {
        struct enic *enic = pmd_priv(eth_dev);
 
+       if (rte_eal_process_type() != RTE_PROC_PRIMARY)
+               return;
+
        ENICPMD_FUNC_TRACE();
        enic->allmulti = 1;
        enic_add_packet_filter(enic);
@@ -496,92 +542,44 @@ static void enicpmd_dev_allmulticast_disable(struct rte_eth_dev *eth_dev)
 {
        struct enic *enic = pmd_priv(eth_dev);
 
+       if (rte_eal_process_type() != RTE_PROC_PRIMARY)
+               return;
+
        ENICPMD_FUNC_TRACE();
        enic->allmulti = 0;
        enic_add_packet_filter(enic);
 }
 
-static void enicpmd_add_mac_addr(struct rte_eth_dev *eth_dev,
+static int enicpmd_add_mac_addr(struct rte_eth_dev *eth_dev,
        struct ether_addr *mac_addr,
        __rte_unused uint32_t index, __rte_unused uint32_t pool)
 {
        struct enic *enic = pmd_priv(eth_dev);
 
+       if (rte_eal_process_type() != RTE_PROC_PRIMARY)
+               return -E_RTE_SECONDARY;
+
        ENICPMD_FUNC_TRACE();
-       enic_set_mac_address(enic, mac_addr->addr_bytes);
+       return enic_set_mac_address(enic, mac_addr->addr_bytes);
 }
 
-static void enicpmd_remove_mac_addr(struct rte_eth_dev *eth_dev, __rte_unused uint32_t index)
+static void enicpmd_remove_mac_addr(struct rte_eth_dev *eth_dev, uint32_t index)
 {
        struct enic *enic = pmd_priv(eth_dev);
 
+       if (rte_eal_process_type() != RTE_PROC_PRIMARY)
+               return;
+
        ENICPMD_FUNC_TRACE();
-       enic_del_mac_address(enic);
-}
-
-
-static uint16_t enicpmd_xmit_pkts(void *tx_queue, struct rte_mbuf **tx_pkts,
-       uint16_t nb_pkts)
-{
-       uint16_t index;
-       unsigned int frags;
-       unsigned int pkt_len;
-       unsigned int seg_len;
-       unsigned int inc_len;
-       unsigned int nb_segs;
-       struct rte_mbuf *tx_pkt, *next_tx_pkt;
-       struct vnic_wq *wq = (struct vnic_wq *)tx_queue;
-       struct enic *enic = vnic_dev_priv(wq->vdev);
-       unsigned short vlan_id;
-       unsigned short ol_flags;
-       uint8_t last_seg, eop;
-       unsigned int host_tx_descs = 0;
-
-       for (index = 0; index < nb_pkts; index++) {
-               tx_pkt = *tx_pkts++;
-               inc_len = 0;
-               nb_segs = tx_pkt->nb_segs;
-               if (nb_segs > vnic_wq_desc_avail(wq)) {
-                       if (index > 0)
-                               enic_post_wq_index(wq);
-
-                       /* wq cleanup and try again */
-                       if (!enic_cleanup_wq(enic, wq) ||
-                               (nb_segs > vnic_wq_desc_avail(wq))) {
-                               return index;
-                       }
-               }
-
-               pkt_len = tx_pkt->pkt_len;
-               vlan_id = tx_pkt->vlan_tci;
-               ol_flags = tx_pkt->ol_flags;
-               for (frags = 0; inc_len < pkt_len; frags++) {
-                       if (!tx_pkt)
-                               break;
-                       next_tx_pkt = tx_pkt->next;
-                       seg_len = tx_pkt->data_len;
-                       inc_len += seg_len;
-
-                       host_tx_descs++;
-                       last_seg = 0;
-                       eop = 0;
-                       if ((pkt_len == inc_len) || !next_tx_pkt) {
-                               eop = 1;
-                               /* post if last packet in batch or > thresh */
-                               if ((index == (nb_pkts - 1)) ||
-                                  (host_tx_descs > ENIC_TX_POST_THRESH)) {
-                                       last_seg = 1;
-                                       host_tx_descs = 0;
-                               }
-                       }
-                       enic_send_pkt(enic, wq, tx_pkt, (unsigned short)seg_len,
-                                     !frags, eop, last_seg, ol_flags, vlan_id);
-                       tx_pkt = next_tx_pkt;
-               }
-       }
+       enic_del_mac_address(enic, index);
+}
+
+static int enicpmd_mtu_set(struct rte_eth_dev *eth_dev, uint16_t mtu)
+{
+       struct enic *enic = pmd_priv(eth_dev);
 
-       enic_cleanup_wq(enic, wq);
-       return index;
+       ENICPMD_FUNC_TRACE();
+       return enic_set_mtu(enic, mtu);
 }
 
 static const struct eth_dev_ops enicpmd_eth_dev_ops = {
@@ -601,7 +599,7 @@ static const struct eth_dev_ops enicpmd_eth_dev_ops = {
        .queue_stats_mapping_set = NULL,
        .dev_infos_get        = enicpmd_dev_info_get,
        .dev_supported_ptypes_get = enicpmd_dev_supported_ptypes_get,
-       .mtu_set              = NULL,
+       .mtu_set              = enicpmd_mtu_set,
        .vlan_filter_set      = enicpmd_vlan_filter_set,
        .vlan_tpid_set        = NULL,
        .vlan_offload_set     = enicpmd_vlan_offload_set,
@@ -612,7 +610,7 @@ static const struct eth_dev_ops enicpmd_eth_dev_ops = {
        .tx_queue_stop        = enicpmd_dev_tx_queue_stop,
        .rx_queue_setup       = enicpmd_dev_rx_queue_setup,
        .rx_queue_release     = enicpmd_dev_rx_queue_release,
-       .rx_queue_count       = NULL,
+       .rx_queue_count       = enicpmd_dev_rx_queue_count,
        .rx_descriptor_done   = NULL,
        .tx_queue_setup       = enicpmd_dev_tx_queue_setup,
        .tx_queue_release     = enicpmd_dev_tx_queue_release,
@@ -642,9 +640,10 @@ static int eth_enicpmd_dev_init(struct rte_eth_dev *eth_dev)
        enic->rte_dev = eth_dev;
        eth_dev->dev_ops = &enicpmd_eth_dev_ops;
        eth_dev->rx_pkt_burst = &enic_recv_pkts;
-       eth_dev->tx_pkt_burst = &enicpmd_xmit_pkts;
+       eth_dev->tx_pkt_burst = &enic_xmit_pkts;
+       eth_dev->tx_pkt_prepare = &enic_prep_pkts;
 
-       pdev = eth_dev->pci_dev;
+       pdev = RTE_ETH_DEV_TO_PCI(eth_dev);
        rte_eth_copy_pci_info(eth_dev, pdev);
        enic->pdev = pdev;
        addr = &pdev->addr;
@@ -655,33 +654,25 @@ static int eth_enicpmd_dev_init(struct rte_eth_dev *eth_dev)
        return enic_probe(enic);
 }
 
-static struct eth_driver rte_enic_pmd = {
-       .pci_drv = {
-               .name = "rte_enic_pmd",
-               .id_table = pci_id_enic_map,
-               .drv_flags = RTE_PCI_DRV_NEED_MAPPING,
-       },
-       .eth_dev_init = eth_enicpmd_dev_init,
-       .dev_private_size = sizeof(struct enic),
-};
-
-/* Driver initialization routine.
- * Invoked once at EAL init time.
- * Register as the [Poll Mode] Driver of Cisco ENIC device.
- */
-static int
-rte_enic_pmd_init(__rte_unused const char *name,
-        __rte_unused const char *params)
+static int eth_enic_pci_probe(struct rte_pci_driver *pci_drv __rte_unused,
+       struct rte_pci_device *pci_dev)
 {
-       ENICPMD_FUNC_TRACE();
+       return rte_eth_dev_pci_generic_probe(pci_dev, sizeof(struct enic),
+               eth_enicpmd_dev_init);
+}
 
-       rte_eth_driver_register(&rte_enic_pmd);
-       return 0;
+static int eth_enic_pci_remove(struct rte_pci_device *pci_dev)
+{
+       return rte_eth_dev_pci_generic_remove(pci_dev, NULL);
 }
 
-static struct rte_driver rte_enic_driver = {
-       .type = PMD_PDEV,
-       .init = rte_enic_pmd_init,
+static struct rte_pci_driver rte_enic_pmd = {
+       .id_table = pci_id_enic_map,
+       .drv_flags = RTE_PCI_DRV_NEED_MAPPING | RTE_PCI_DRV_INTR_LSC,
+       .probe = eth_enic_pci_probe,
+       .remove = eth_enic_pci_remove,
 };
 
-PMD_REGISTER_DRIVER(rte_enic_driver);
+RTE_PMD_REGISTER_PCI(net_enic, rte_enic_pmd);
+RTE_PMD_REGISTER_PCI_TABLE(net_enic, pci_id_enic_map);
+RTE_PMD_REGISTER_KMOD_DEP(net_enic, "* igb_uio | uio_pci_generic | vfio-pci");