Imported Upstream version 16.04
[deb_dpdk.git] / drivers / net / ixgbe / base / ixgbe_dcb_82599.c
diff --git a/drivers/net/ixgbe/base/ixgbe_dcb_82599.c b/drivers/net/ixgbe/base/ixgbe_dcb_82599.c
new file mode 100644 (file)
index 0000000..a52f83a
--- /dev/null
@@ -0,0 +1,593 @@
+/*******************************************************************************
+
+Copyright (c) 2001-2015, Intel Corporation
+All rights reserved.
+
+Redistribution and use in source and binary forms, with or without
+modification, are permitted provided that the following conditions are met:
+
+ 1. Redistributions of source code must retain the above copyright notice,
+    this list of conditions and the following disclaimer.
+
+ 2. Redistributions in binary form must reproduce the above copyright
+    notice, this list of conditions and the following disclaimer in the
+    documentation and/or other materials provided with the distribution.
+
+ 3. Neither the name of the Intel Corporation nor the names of its
+    contributors may be used to endorse or promote products derived from
+    this software without specific prior written permission.
+
+THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
+AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
+IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
+ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE
+LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
+CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
+SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
+INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
+CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
+ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
+POSSIBILITY OF SUCH DAMAGE.
+
+***************************************************************************/
+
+
+#include "ixgbe_type.h"
+#include "ixgbe_dcb.h"
+#include "ixgbe_dcb_82599.h"
+
+/**
+ * ixgbe_dcb_get_tc_stats_82599 - Returns status for each traffic class
+ * @hw: pointer to hardware structure
+ * @stats: pointer to statistics structure
+ * @tc_count:  Number of elements in bwg_array.
+ *
+ * This function returns the status data for each of the Traffic Classes in use.
+ */
+s32 ixgbe_dcb_get_tc_stats_82599(struct ixgbe_hw *hw,
+                                struct ixgbe_hw_stats *stats,
+                                u8 tc_count)
+{
+       int tc;
+
+       DEBUGFUNC("dcb_get_tc_stats");
+
+       if (tc_count > IXGBE_DCB_MAX_TRAFFIC_CLASS)
+               return IXGBE_ERR_PARAM;
+
+       /* Statistics pertaining to each traffic class */
+       for (tc = 0; tc < tc_count; tc++) {
+               /* Transmitted Packets */
+               stats->qptc[tc] += IXGBE_READ_REG(hw, IXGBE_QPTC(tc));
+               /* Transmitted Bytes (read low first to prevent missed carry) */
+               stats->qbtc[tc] += IXGBE_READ_REG(hw, IXGBE_QBTC_L(tc));
+               stats->qbtc[tc] +=
+                       (((u64)(IXGBE_READ_REG(hw, IXGBE_QBTC_H(tc)))) << 32);
+               /* Received Packets */
+               stats->qprc[tc] += IXGBE_READ_REG(hw, IXGBE_QPRC(tc));
+               /* Received Bytes (read low first to prevent missed carry) */
+               stats->qbrc[tc] += IXGBE_READ_REG(hw, IXGBE_QBRC_L(tc));
+               stats->qbrc[tc] +=
+                       (((u64)(IXGBE_READ_REG(hw, IXGBE_QBRC_H(tc)))) << 32);
+
+               /* Received Dropped Packet */
+               stats->qprdc[tc] += IXGBE_READ_REG(hw, IXGBE_QPRDC(tc));
+       }
+
+       return IXGBE_SUCCESS;
+}
+
+/**
+ * ixgbe_dcb_get_pfc_stats_82599 - Return CBFC status data
+ * @hw: pointer to hardware structure
+ * @stats: pointer to statistics structure
+ * @tc_count:  Number of elements in bwg_array.
+ *
+ * This function returns the CBFC status data for each of the Traffic Classes.
+ */
+s32 ixgbe_dcb_get_pfc_stats_82599(struct ixgbe_hw *hw,
+                                 struct ixgbe_hw_stats *stats,
+                                 u8 tc_count)
+{
+       int tc;
+
+       DEBUGFUNC("dcb_get_pfc_stats");
+
+       if (tc_count > IXGBE_DCB_MAX_TRAFFIC_CLASS)
+               return IXGBE_ERR_PARAM;
+
+       for (tc = 0; tc < tc_count; tc++) {
+               /* Priority XOFF Transmitted */
+               stats->pxofftxc[tc] += IXGBE_READ_REG(hw, IXGBE_PXOFFTXC(tc));
+               /* Priority XOFF Received */
+               stats->pxoffrxc[tc] += IXGBE_READ_REG(hw, IXGBE_PXOFFRXCNT(tc));
+       }
+
+       return IXGBE_SUCCESS;
+}
+
+/**
+ * ixgbe_dcb_config_rx_arbiter_82599 - Config Rx Data arbiter
+ * @hw: pointer to hardware structure
+ * @dcb_config: pointer to ixgbe_dcb_config structure
+ *
+ * Configure Rx Packet Arbiter and credits for each traffic class.
+ */
+s32 ixgbe_dcb_config_rx_arbiter_82599(struct ixgbe_hw *hw, u16 *refill,
+                                     u16 *max, u8 *bwg_id, u8 *tsa,
+                                     u8 *map)
+{
+       u32 reg = 0;
+       u32 credit_refill = 0;
+       u32 credit_max = 0;
+       u8  i = 0;
+
+       /*
+        * Disable the arbiter before changing parameters
+        * (always enable recycle mode; WSP)
+        */
+       reg = IXGBE_RTRPCS_RRM | IXGBE_RTRPCS_RAC | IXGBE_RTRPCS_ARBDIS;
+       IXGBE_WRITE_REG(hw, IXGBE_RTRPCS, reg);
+
+       /*
+        * map all UPs to TCs. up_to_tc_bitmap for each TC has corresponding
+        * bits sets for the UPs that needs to be mappped to that TC.
+        * e.g if priorities 6 and 7 are to be mapped to a TC then the
+        * up_to_tc_bitmap value for that TC will be 11000000 in binary.
+        */
+       reg = 0;
+       for (i = 0; i < IXGBE_DCB_MAX_USER_PRIORITY; i++)
+               reg |= (map[i] << (i * IXGBE_RTRUP2TC_UP_SHIFT));
+
+       IXGBE_WRITE_REG(hw, IXGBE_RTRUP2TC, reg);
+
+       /* Configure traffic class credits and priority */
+       for (i = 0; i < IXGBE_DCB_MAX_TRAFFIC_CLASS; i++) {
+               credit_refill = refill[i];
+               credit_max = max[i];
+               reg = credit_refill | (credit_max << IXGBE_RTRPT4C_MCL_SHIFT);
+
+               reg |= (u32)(bwg_id[i]) << IXGBE_RTRPT4C_BWG_SHIFT;
+
+               if (tsa[i] == ixgbe_dcb_tsa_strict)
+                       reg |= IXGBE_RTRPT4C_LSP;
+
+               IXGBE_WRITE_REG(hw, IXGBE_RTRPT4C(i), reg);
+       }
+
+       /*
+        * Configure Rx packet plane (recycle mode; WSP) and
+        * enable arbiter
+        */
+       reg = IXGBE_RTRPCS_RRM | IXGBE_RTRPCS_RAC;
+       IXGBE_WRITE_REG(hw, IXGBE_RTRPCS, reg);
+
+       return IXGBE_SUCCESS;
+}
+
+/**
+ * ixgbe_dcb_config_tx_desc_arbiter_82599 - Config Tx Desc. arbiter
+ * @hw: pointer to hardware structure
+ * @dcb_config: pointer to ixgbe_dcb_config structure
+ *
+ * Configure Tx Descriptor Arbiter and credits for each traffic class.
+ */
+s32 ixgbe_dcb_config_tx_desc_arbiter_82599(struct ixgbe_hw *hw, u16 *refill,
+                                          u16 *max, u8 *bwg_id, u8 *tsa)
+{
+       u32 reg, max_credits;
+       u8  i;
+
+       /* Clear the per-Tx queue credits; we use per-TC instead */
+       for (i = 0; i < 128; i++) {
+               IXGBE_WRITE_REG(hw, IXGBE_RTTDQSEL, i);
+               IXGBE_WRITE_REG(hw, IXGBE_RTTDT1C, 0);
+       }
+
+       /* Configure traffic class credits and priority */
+       for (i = 0; i < IXGBE_DCB_MAX_TRAFFIC_CLASS; i++) {
+               max_credits = max[i];
+               reg = max_credits << IXGBE_RTTDT2C_MCL_SHIFT;
+               reg |= refill[i];
+               reg |= (u32)(bwg_id[i]) << IXGBE_RTTDT2C_BWG_SHIFT;
+
+               if (tsa[i] == ixgbe_dcb_tsa_group_strict_cee)
+                       reg |= IXGBE_RTTDT2C_GSP;
+
+               if (tsa[i] == ixgbe_dcb_tsa_strict)
+                       reg |= IXGBE_RTTDT2C_LSP;
+
+               IXGBE_WRITE_REG(hw, IXGBE_RTTDT2C(i), reg);
+       }
+
+       /*
+        * Configure Tx descriptor plane (recycle mode; WSP) and
+        * enable arbiter
+        */
+       reg = IXGBE_RTTDCS_TDPAC | IXGBE_RTTDCS_TDRM;
+       IXGBE_WRITE_REG(hw, IXGBE_RTTDCS, reg);
+
+       return IXGBE_SUCCESS;
+}
+
+/**
+ * ixgbe_dcb_config_tx_data_arbiter_82599 - Config Tx Data arbiter
+ * @hw: pointer to hardware structure
+ * @dcb_config: pointer to ixgbe_dcb_config structure
+ *
+ * Configure Tx Packet Arbiter and credits for each traffic class.
+ */
+s32 ixgbe_dcb_config_tx_data_arbiter_82599(struct ixgbe_hw *hw, u16 *refill,
+                                          u16 *max, u8 *bwg_id, u8 *tsa,
+                                          u8 *map)
+{
+       u32 reg;
+       u8 i;
+
+       /*
+        * Disable the arbiter before changing parameters
+        * (always enable recycle mode; SP; arb delay)
+        */
+       reg = IXGBE_RTTPCS_TPPAC | IXGBE_RTTPCS_TPRM |
+             (IXGBE_RTTPCS_ARBD_DCB << IXGBE_RTTPCS_ARBD_SHIFT) |
+             IXGBE_RTTPCS_ARBDIS;
+       IXGBE_WRITE_REG(hw, IXGBE_RTTPCS, reg);
+
+       /*
+        * map all UPs to TCs. up_to_tc_bitmap for each TC has corresponding
+        * bits sets for the UPs that needs to be mappped to that TC.
+        * e.g if priorities 6 and 7 are to be mapped to a TC then the
+        * up_to_tc_bitmap value for that TC will be 11000000 in binary.
+        */
+       reg = 0;
+       for (i = 0; i < IXGBE_DCB_MAX_USER_PRIORITY; i++)
+               reg |= (map[i] << (i * IXGBE_RTTUP2TC_UP_SHIFT));
+
+       IXGBE_WRITE_REG(hw, IXGBE_RTTUP2TC, reg);
+
+       /* Configure traffic class credits and priority */
+       for (i = 0; i < IXGBE_DCB_MAX_TRAFFIC_CLASS; i++) {
+               reg = refill[i];
+               reg |= (u32)(max[i]) << IXGBE_RTTPT2C_MCL_SHIFT;
+               reg |= (u32)(bwg_id[i]) << IXGBE_RTTPT2C_BWG_SHIFT;
+
+               if (tsa[i] == ixgbe_dcb_tsa_group_strict_cee)
+                       reg |= IXGBE_RTTPT2C_GSP;
+
+               if (tsa[i] == ixgbe_dcb_tsa_strict)
+                       reg |= IXGBE_RTTPT2C_LSP;
+
+               IXGBE_WRITE_REG(hw, IXGBE_RTTPT2C(i), reg);
+       }
+
+       /*
+        * Configure Tx packet plane (recycle mode; SP; arb delay) and
+        * enable arbiter
+        */
+       reg = IXGBE_RTTPCS_TPPAC | IXGBE_RTTPCS_TPRM |
+             (IXGBE_RTTPCS_ARBD_DCB << IXGBE_RTTPCS_ARBD_SHIFT);
+       IXGBE_WRITE_REG(hw, IXGBE_RTTPCS, reg);
+
+       return IXGBE_SUCCESS;
+}
+
+/**
+ * ixgbe_dcb_config_pfc_82599 - Configure priority flow control
+ * @hw: pointer to hardware structure
+ * @pfc_en: enabled pfc bitmask
+ * @map: priority to tc assignments indexed by priority
+ *
+ * Configure Priority Flow Control (PFC) for each traffic class.
+ */
+s32 ixgbe_dcb_config_pfc_82599(struct ixgbe_hw *hw, u8 pfc_en, u8 *map)
+{
+       u32 i, j, fcrtl, reg;
+       u8 max_tc = 0;
+
+       /* Enable Transmit Priority Flow Control */
+       IXGBE_WRITE_REG(hw, IXGBE_FCCFG, IXGBE_FCCFG_TFCE_PRIORITY);
+
+       /* Enable Receive Priority Flow Control */
+       reg = IXGBE_READ_REG(hw, IXGBE_MFLCN);
+       reg |= IXGBE_MFLCN_DPF;
+
+       /*
+        * X540 supports per TC Rx priority flow control.  So
+        * clear all TCs and only enable those that should be
+        * enabled.
+        */
+       reg &= ~(IXGBE_MFLCN_RPFCE_MASK | IXGBE_MFLCN_RFCE);
+
+       if (hw->mac.type >= ixgbe_mac_X540)
+               reg |= pfc_en << IXGBE_MFLCN_RPFCE_SHIFT;
+
+       if (pfc_en)
+               reg |= IXGBE_MFLCN_RPFCE;
+
+       IXGBE_WRITE_REG(hw, IXGBE_MFLCN, reg);
+
+       for (i = 0; i < IXGBE_DCB_MAX_USER_PRIORITY; i++) {
+               if (map[i] > max_tc)
+                       max_tc = map[i];
+       }
+
+
+       /* Configure PFC Tx thresholds per TC */
+       for (i = 0; i <= max_tc; i++) {
+               int enabled = 0;
+
+               for (j = 0; j < IXGBE_DCB_MAX_USER_PRIORITY; j++) {
+                       if ((map[j] == i) && (pfc_en & (1 << j))) {
+                               enabled = 1;
+                               break;
+                       }
+               }
+
+               if (enabled) {
+                       reg = (hw->fc.high_water[i] << 10) | IXGBE_FCRTH_FCEN;
+                       fcrtl = (hw->fc.low_water[i] << 10) | IXGBE_FCRTL_XONE;
+                       IXGBE_WRITE_REG(hw, IXGBE_FCRTL_82599(i), fcrtl);
+               } else {
+                       /*
+                        * In order to prevent Tx hangs when the internal Tx
+                        * switch is enabled we must set the high water mark
+                        * to the Rx packet buffer size - 24KB.  This allows
+                        * the Tx switch to function even under heavy Rx
+                        * workloads.
+                        */
+                       reg = IXGBE_READ_REG(hw, IXGBE_RXPBSIZE(i)) - 24576;
+                       IXGBE_WRITE_REG(hw, IXGBE_FCRTL_82599(i), 0);
+               }
+
+               IXGBE_WRITE_REG(hw, IXGBE_FCRTH_82599(i), reg);
+       }
+
+       for (; i < IXGBE_DCB_MAX_TRAFFIC_CLASS; i++) {
+               IXGBE_WRITE_REG(hw, IXGBE_FCRTL_82599(i), 0);
+               IXGBE_WRITE_REG(hw, IXGBE_FCRTH_82599(i), 0);
+       }
+
+       /* Configure pause time (2 TCs per register) */
+       reg = hw->fc.pause_time | (hw->fc.pause_time << 16);
+       for (i = 0; i < (IXGBE_DCB_MAX_TRAFFIC_CLASS / 2); i++)
+               IXGBE_WRITE_REG(hw, IXGBE_FCTTV(i), reg);
+
+       /* Configure flow control refresh threshold value */
+       IXGBE_WRITE_REG(hw, IXGBE_FCRTV, hw->fc.pause_time / 2);
+
+       return IXGBE_SUCCESS;
+}
+
+/**
+ * ixgbe_dcb_config_tc_stats_82599 - Config traffic class statistics
+ * @hw: pointer to hardware structure
+ *
+ * Configure queue statistics registers, all queues belonging to same traffic
+ * class uses a single set of queue statistics counters.
+ */
+s32 ixgbe_dcb_config_tc_stats_82599(struct ixgbe_hw *hw,
+                                   struct ixgbe_dcb_config *dcb_config)
+{
+       u32 reg = 0;
+       u8  i   = 0;
+       u8 tc_count = 8;
+       bool vt_mode = false;
+
+       if (dcb_config != NULL) {
+               tc_count = dcb_config->num_tcs.pg_tcs;
+               vt_mode = dcb_config->vt_mode;
+       }
+
+       if (!((tc_count == 8 && vt_mode == false) || tc_count == 4))
+               return IXGBE_ERR_PARAM;
+
+       if (tc_count == 8 && vt_mode == false) {
+               /*
+                * Receive Queues stats setting
+                * 32 RQSMR registers, each configuring 4 queues.
+                *
+                * Set all 16 queues of each TC to the same stat
+                * with TC 'n' going to stat 'n'.
+                */
+               for (i = 0; i < 32; i++) {
+                       reg = 0x01010101 * (i / 4);
+                       IXGBE_WRITE_REG(hw, IXGBE_RQSMR(i), reg);
+               }
+               /*
+                * Transmit Queues stats setting
+                * 32 TQSM registers, each controlling 4 queues.
+                *
+                * Set all queues of each TC to the same stat
+                * with TC 'n' going to stat 'n'.
+                * Tx queues are allocated non-uniformly to TCs:
+                * 32, 32, 16, 16, 8, 8, 8, 8.
+                */
+               for (i = 0; i < 32; i++) {
+                       if (i < 8)
+                               reg = 0x00000000;
+                       else if (i < 16)
+                               reg = 0x01010101;
+                       else if (i < 20)
+                               reg = 0x02020202;
+                       else if (i < 24)
+                               reg = 0x03030303;
+                       else if (i < 26)
+                               reg = 0x04040404;
+                       else if (i < 28)
+                               reg = 0x05050505;
+                       else if (i < 30)
+                               reg = 0x06060606;
+                       else
+                               reg = 0x07070707;
+                       IXGBE_WRITE_REG(hw, IXGBE_TQSM(i), reg);
+               }
+       } else if (tc_count == 4 && vt_mode == false) {
+               /*
+                * Receive Queues stats setting
+                * 32 RQSMR registers, each configuring 4 queues.
+                *
+                * Set all 16 queues of each TC to the same stat
+                * with TC 'n' going to stat 'n'.
+                */
+               for (i = 0; i < 32; i++) {
+                       if (i % 8 > 3)
+                               /* In 4 TC mode, odd 16-queue ranges are
+                                *  not used.
+                               */
+                               continue;
+                       reg = 0x01010101 * (i / 8);
+                       IXGBE_WRITE_REG(hw, IXGBE_RQSMR(i), reg);
+               }
+               /*
+                * Transmit Queues stats setting
+                * 32 TQSM registers, each controlling 4 queues.
+                *
+                * Set all queues of each TC to the same stat
+                * with TC 'n' going to stat 'n'.
+                * Tx queues are allocated non-uniformly to TCs:
+                * 64, 32, 16, 16.
+                */
+               for (i = 0; i < 32; i++) {
+                       if (i < 16)
+                               reg = 0x00000000;
+                       else if (i < 24)
+                               reg = 0x01010101;
+                       else if (i < 28)
+                               reg = 0x02020202;
+                       else
+                               reg = 0x03030303;
+                       IXGBE_WRITE_REG(hw, IXGBE_TQSM(i), reg);
+               }
+       } else if (tc_count == 4 && vt_mode == true) {
+               /*
+                * Receive Queues stats setting
+                * 32 RQSMR registers, each configuring 4 queues.
+                *
+                * Queue Indexing in 32 VF with DCB mode maps 4 TC's to each
+                * pool. Set all 32 queues of each TC across pools to the same
+                * stat with TC 'n' going to stat 'n'.
+                */
+               for (i = 0; i < 32; i++)
+                       IXGBE_WRITE_REG(hw, IXGBE_RQSMR(i), 0x03020100);
+               /*
+                * Transmit Queues stats setting
+                * 32 TQSM registers, each controlling 4 queues.
+                *
+                * Queue Indexing in 32 VF with DCB mode maps 4 TC's to each
+                * pool. Set all 32 queues of each TC across pools to the same
+                * stat with TC 'n' going to stat 'n'.
+                */
+               for (i = 0; i < 32; i++)
+                       IXGBE_WRITE_REG(hw, IXGBE_TQSM(i), 0x03020100);
+       }
+
+       return IXGBE_SUCCESS;
+}
+
+/**
+ * ixgbe_dcb_config_82599 - Configure general DCB parameters
+ * @hw: pointer to hardware structure
+ * @dcb_config: pointer to ixgbe_dcb_config structure
+ *
+ * Configure general DCB parameters.
+ */
+s32 ixgbe_dcb_config_82599(struct ixgbe_hw *hw,
+                          struct ixgbe_dcb_config *dcb_config)
+{
+       u32 reg;
+       u32 q;
+
+       /* Disable the Tx desc arbiter so that MTQC can be changed */
+       reg = IXGBE_READ_REG(hw, IXGBE_RTTDCS);
+       reg |= IXGBE_RTTDCS_ARBDIS;
+       IXGBE_WRITE_REG(hw, IXGBE_RTTDCS, reg);
+
+       reg = IXGBE_READ_REG(hw, IXGBE_MRQC);
+       if (dcb_config->num_tcs.pg_tcs == 8) {
+               /* Enable DCB for Rx with 8 TCs */
+               switch (reg & IXGBE_MRQC_MRQE_MASK) {
+               case 0:
+               case IXGBE_MRQC_RT4TCEN:
+                       /* RSS disabled cases */
+                       reg = (reg & ~IXGBE_MRQC_MRQE_MASK) |
+                             IXGBE_MRQC_RT8TCEN;
+                       break;
+               case IXGBE_MRQC_RSSEN:
+               case IXGBE_MRQC_RTRSS4TCEN:
+                       /* RSS enabled cases */
+                       reg = (reg & ~IXGBE_MRQC_MRQE_MASK) |
+                             IXGBE_MRQC_RTRSS8TCEN;
+                       break;
+               default:
+                       /*
+                        * Unsupported value, assume stale data,
+                        * overwrite no RSS
+                        */
+                       ASSERT(0);
+                       reg = (reg & ~IXGBE_MRQC_MRQE_MASK) |
+                             IXGBE_MRQC_RT8TCEN;
+               }
+       }
+       if (dcb_config->num_tcs.pg_tcs == 4) {
+               /* We support both VT-on and VT-off with 4 TCs. */
+               if (dcb_config->vt_mode)
+                       reg = (reg & ~IXGBE_MRQC_MRQE_MASK) |
+                             IXGBE_MRQC_VMDQRT4TCEN;
+               else
+                       reg = (reg & ~IXGBE_MRQC_MRQE_MASK) |
+                             IXGBE_MRQC_RTRSS4TCEN;
+       }
+       IXGBE_WRITE_REG(hw, IXGBE_MRQC, reg);
+
+       /* Enable DCB for Tx with 8 TCs */
+       if (dcb_config->num_tcs.pg_tcs == 8)
+               reg = IXGBE_MTQC_RT_ENA | IXGBE_MTQC_8TC_8TQ;
+       else {
+               /* We support both VT-on and VT-off with 4 TCs. */
+               reg = IXGBE_MTQC_RT_ENA | IXGBE_MTQC_4TC_4TQ;
+               if (dcb_config->vt_mode)
+                       reg |= IXGBE_MTQC_VT_ENA;
+       }
+       IXGBE_WRITE_REG(hw, IXGBE_MTQC, reg);
+
+       /* Disable drop for all queues */
+       for (q = 0; q < 128; q++)
+               IXGBE_WRITE_REG(hw, IXGBE_QDE,
+                               (IXGBE_QDE_WRITE | (q << IXGBE_QDE_IDX_SHIFT)));
+
+       /* Enable the Tx desc arbiter */
+       reg = IXGBE_READ_REG(hw, IXGBE_RTTDCS);
+       reg &= ~IXGBE_RTTDCS_ARBDIS;
+       IXGBE_WRITE_REG(hw, IXGBE_RTTDCS, reg);
+
+       /* Enable Security TX Buffer IFG for DCB */
+       reg = IXGBE_READ_REG(hw, IXGBE_SECTXMINIFG);
+       reg |= IXGBE_SECTX_DCB;
+       IXGBE_WRITE_REG(hw, IXGBE_SECTXMINIFG, reg);
+
+       return IXGBE_SUCCESS;
+}
+
+/**
+ * ixgbe_dcb_hw_config_82599 - Configure and enable DCB
+ * @hw: pointer to hardware structure
+ * @dcb_config: pointer to ixgbe_dcb_config structure
+ *
+ * Configure dcb settings and enable dcb mode.
+ */
+s32 ixgbe_dcb_hw_config_82599(struct ixgbe_hw *hw, int link_speed,
+                             u16 *refill, u16 *max, u8 *bwg_id, u8 *tsa,
+                             u8 *map)
+{
+       UNREFERENCED_1PARAMETER(link_speed);
+
+       ixgbe_dcb_config_rx_arbiter_82599(hw, refill, max, bwg_id, tsa,
+                                         map);
+       ixgbe_dcb_config_tx_desc_arbiter_82599(hw, refill, max, bwg_id,
+                                              tsa);
+       ixgbe_dcb_config_tx_data_arbiter_82599(hw, refill, max, bwg_id,
+                                              tsa, map);
+
+       return IXGBE_SUCCESS;
+}
+