misc: deprecate dpdk hqos
[vpp.git] / src / plugins / dpdk / hqos / hqos.c
diff --git a/src/plugins/dpdk/hqos/hqos.c b/src/plugins/dpdk/hqos/hqos.c
deleted file mode 100644 (file)
index 1a8dd6d..0000000
+++ /dev/null
@@ -1,771 +0,0 @@
-/*
- * Copyright(c) 2016 Intel Corporation. All rights reserved.
- * Licensed under the Apache License, Version 2.0 (the "License");
- * you may not use this file except in compliance with the License.
- * You may obtain a copy of the License at:
- *
- *     http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-#include <stdio.h>
-#include <stdlib.h>
-#include <unistd.h>
-#include <sys/stat.h>
-#include <sys/mount.h>
-#include <string.h>
-#include <fcntl.h>
-
-#include <vppinfra/vec.h>
-#include <vppinfra/error.h>
-#include <vppinfra/format.h>
-#include <vppinfra/bitmap.h>
-
-#include <vnet/vnet.h>
-#include <vnet/ethernet/ethernet.h>
-#include <dpdk/device/dpdk.h>
-
-#include <vlib/pci/pci.h>
-#include <vlibmemory/api.h>
-#include <vlibmemory/vl_memory_msg_enum.h>     /* enumerate all vlib messages */
-
-#define vl_typedefs            /* define message structures */
-#include <vlibmemory/vl_memory_api_h.h>
-#undef vl_typedefs
-
-/* instantiate all the print functions we know about */
-#define vl_print(handle, ...) vlib_cli_output (handle, __VA_ARGS__)
-#define vl_printfun
-#include <vlibmemory/vl_memory_api_h.h>
-#undef vl_printfun
-
-#include <dpdk/device/dpdk_priv.h>
-
-/***
- *
- * HQoS default configuration values
- *
- ***/
-
-static dpdk_device_config_hqos_t hqos_params_default = {
-  .hqos_thread_valid = 0,
-
-  .swq_size = 4096,
-  .burst_enq = 256,
-  .burst_deq = 220,
-
-  /*
-   * Packet field to identify the subport.
-   *
-   * Default value: Since only one subport is defined by default (see below:
-   *     n_subports_per_port = 1), the subport ID is hardcoded to 0.
-   */
-  .pktfield0_slabpos = 0,
-  .pktfield0_slabmask = 0,
-
-  /*
-   * Packet field to identify the pipe.
-   *
-   * Default value: Assuming Ethernet/IPv4/UDP packets, UDP payload bits 12 .. 23
-   */
-  .pktfield1_slabpos = 40,
-  .pktfield1_slabmask = 0x0000000FFF000000LLU,
-
-  /* Packet field used as index into TC translation table to identify the traffic
-   *     class and queue.
-   *
-   * Default value: Assuming Ethernet/IPv4 packets, IPv4 DSCP field
-   */
-  .pktfield2_slabpos = 8,
-  .pktfield2_slabmask = 0x00000000000000FCLLU,
-  .tc_table = {
-              0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15,
-              0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15,
-              0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15,
-              0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15,
-              },
-
-  /* port */
-  .port = {
-          .name = NULL,        /* Set at init */
-          .socket = 0,         /* Set at init */
-          .rate = 1250000000,  /* Assuming 10GbE port */
-          .mtu = 14 + 1500,    /* Assuming Ethernet/IPv4 pkt (Ethernet FCS not included) */
-          .frame_overhead = RTE_SCHED_FRAME_OVERHEAD_DEFAULT,
-          .n_subports_per_port = 1,
-          .n_pipes_per_subport = 4096,
-          .qsize = {64, 64, 64, 64},
-          .pipe_profiles = NULL,       /* Set at config */
-          .n_pipe_profiles = 1,
-
-#ifdef RTE_SCHED_RED
-          .red_params = {
-                         /* Traffic Class 0 Colors Green / Yellow / Red */
-                         [0][0] = {.min_th = 48,.max_th = 64,.maxp_inv =
-                                   10,.wq_log2 = 9},
-                         [0][1] = {.min_th = 40,.max_th = 64,.maxp_inv =
-                                   10,.wq_log2 = 9},
-                         [0][2] = {.min_th = 32,.max_th = 64,.maxp_inv =
-                                   10,.wq_log2 = 9},
-
-                         /* Traffic Class 1 - Colors Green / Yellow / Red */
-                         [1][0] = {.min_th = 48,.max_th = 64,.maxp_inv =
-                                   10,.wq_log2 = 9},
-                         [1][1] = {.min_th = 40,.max_th = 64,.maxp_inv =
-                                   10,.wq_log2 = 9},
-                         [1][2] = {.min_th = 32,.max_th = 64,.maxp_inv =
-                                   10,.wq_log2 = 9},
-
-                         /* Traffic Class 2 - Colors Green / Yellow / Red */
-                         [2][0] = {.min_th = 48,.max_th = 64,.maxp_inv =
-                                   10,.wq_log2 = 9},
-                         [2][1] = {.min_th = 40,.max_th = 64,.maxp_inv =
-                                   10,.wq_log2 = 9},
-                         [2][2] = {.min_th = 32,.max_th = 64,.maxp_inv =
-                                   10,.wq_log2 = 9},
-
-                         /* Traffic Class 3 - Colors Green / Yellow / Red */
-                         [3][0] = {.min_th = 48,.max_th = 64,.maxp_inv =
-                                   10,.wq_log2 = 9},
-                         [3][1] = {.min_th = 40,.max_th = 64,.maxp_inv =
-                                   10,.wq_log2 = 9},
-                         [3][2] = {.min_th = 32,.max_th = 64,.maxp_inv =
-                                   10,.wq_log2 = 9}
-                         },
-#endif /* RTE_SCHED_RED */
-          },
-};
-
-static struct rte_sched_subport_params hqos_subport_params_default = {
-  .tb_rate = 1250000000,       /* 10GbE line rate (measured in bytes/second) */
-  .tb_size = 1000000,
-  .tc_rate = {1250000000, 1250000000, 1250000000, 1250000000},
-  .tc_period = 10,
-};
-
-static struct rte_sched_pipe_params hqos_pipe_params_default = {
-  .tb_rate = 305175,           /* 10GbE line rate divided by 4K pipes */
-  .tb_size = 1000000,
-  .tc_rate = {305175, 305175, 305175, 305175},
-  .tc_period = 40,
-#ifdef RTE_SCHED_SUBPORT_TC_OV
-  .tc_ov_weight = 1,
-#endif
-  .wrr_weights = {1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1},
-};
-
-/***
- *
- * HQoS configuration
- *
- ***/
-
-int
-dpdk_hqos_validate_mask (u64 mask, u32 n)
-{
-  int count = __builtin_popcountll (mask);
-  int pos_lead = sizeof (u64) * 8 - count_leading_zeros (mask);
-  int pos_trail = count_trailing_zeros (mask);
-  int count_expected = __builtin_popcount (n - 1);
-
-  /* Handle the exceptions */
-  if (n == 0)
-    return -1;                 /* Error */
-
-  if ((mask == 0) && (n == 1))
-    return 0;                  /* OK */
-
-  if (((mask == 0) && (n != 1)) || ((mask != 0) && (n == 1)))
-    return -2;                 /* Error */
-
-  /* Check that mask is contiguous */
-  if ((pos_lead - pos_trail) != count)
-    return -3;                 /* Error */
-
-  /* Check that mask contains the expected number of bits set */
-  if (count != count_expected)
-    return -4;                 /* Error */
-
-  return 0;                    /* OK */
-}
-
-void
-dpdk_device_config_hqos_pipe_profile_default (dpdk_device_config_hqos_t *
-                                             hqos, u32 pipe_profile_id)
-{
-  memcpy (&hqos->pipe[pipe_profile_id], &hqos_pipe_params_default,
-         sizeof (hqos_pipe_params_default));
-}
-
-void
-dpdk_device_config_hqos_default (dpdk_device_config_hqos_t * hqos)
-{
-  struct rte_sched_subport_params *subport_params;
-  struct rte_sched_pipe_params *pipe_params;
-  u32 *pipe_map;
-  u32 i;
-
-  memcpy (hqos, &hqos_params_default, sizeof (hqos_params_default));
-
-  /* pipe */
-  vec_add2 (hqos->pipe, pipe_params, hqos->port.n_pipe_profiles);
-
-  for (i = 0; i < vec_len (hqos->pipe); i++)
-    memcpy (&pipe_params[i],
-           &hqos_pipe_params_default, sizeof (hqos_pipe_params_default));
-
-  hqos->port.pipe_profiles = hqos->pipe;
-
-  /* subport */
-  vec_add2 (hqos->subport, subport_params, hqos->port.n_subports_per_port);
-
-  for (i = 0; i < vec_len (hqos->subport); i++)
-    memcpy (&subport_params[i],
-           &hqos_subport_params_default,
-           sizeof (hqos_subport_params_default));
-
-  /* pipe profile */
-  vec_add2 (hqos->pipe_map,
-           pipe_map,
-           hqos->port.n_subports_per_port * hqos->port.n_pipes_per_subport);
-
-  for (i = 0; i < vec_len (hqos->pipe_map); i++)
-    pipe_map[i] = 0;
-}
-
-/***
- *
- * HQoS init
- *
- ***/
-
-clib_error_t *
-dpdk_port_setup_hqos (dpdk_device_t * xd, dpdk_device_config_hqos_t * hqos)
-{
-  vlib_thread_main_t *tm = vlib_get_thread_main ();
-  char name[32];
-  u32 subport_id, i;
-  int rv;
-
-  /* Detect the set of worker threads */
-  int worker_thread_first = 0;
-  int worker_thread_count = 0;
-
-  uword *p = hash_get_mem (tm->thread_registrations_by_name, "workers");
-  vlib_thread_registration_t *tr =
-    p ? (vlib_thread_registration_t *) p[0] : 0;
-
-  if (tr && tr->count > 0)
-    {
-      worker_thread_first = tr->first_index;
-      worker_thread_count = tr->count;
-    }
-
-  /* Allocate the per-thread device data array */
-  vec_validate_aligned (xd->hqos_wt, tm->n_vlib_mains - 1,
-                       CLIB_CACHE_LINE_BYTES);
-  clib_memset (xd->hqos_wt, 0, tm->n_vlib_mains * sizeof (xd->hqos_wt[0]));
-
-  vec_validate_aligned (xd->hqos_ht, 0, CLIB_CACHE_LINE_BYTES);
-  clib_memset (xd->hqos_ht, 0, sizeof (xd->hqos_ht[0]));
-
-  /* Allocate space for one SWQ per worker thread in the I/O TX thread data structure */
-  vec_validate (xd->hqos_ht->swq, worker_thread_count);
-
-  /* SWQ */
-  for (i = 0; i < worker_thread_count + 1; i++)
-    {
-      u32 swq_flags = RING_F_SP_ENQ | RING_F_SC_DEQ;
-
-      snprintf (name, sizeof (name), "SWQ-worker%u-to-device%u", i,
-               xd->port_id);
-      xd->hqos_ht->swq[i] =
-       rte_ring_create (name, hqos->swq_size, xd->cpu_socket, swq_flags);
-      if (xd->hqos_ht->swq[i] == NULL)
-       return clib_error_return (0,
-                                 "SWQ-worker%u-to-device%u: rte_ring_create err",
-                                 i, xd->port_id);
-    }
-
-  /*
-   * HQoS
-   */
-
-  /* HQoS port */
-  snprintf (name, sizeof (name), "HQoS%u", xd->port_id);
-  hqos->port.name = strdup (name);
-  if (hqos->port.name == NULL)
-    return clib_error_return (0, "HQoS%u: strdup err", xd->port_id);
-
-  hqos->port.socket = rte_eth_dev_socket_id (xd->port_id);
-  if (hqos->port.socket == SOCKET_ID_ANY)
-    hqos->port.socket = 0;
-
-  xd->hqos_ht->hqos = rte_sched_port_config (&hqos->port);
-  if (xd->hqos_ht->hqos == NULL)
-    return clib_error_return (0, "HQoS%u: rte_sched_port_config err",
-                             xd->port_id);
-
-  /* HQoS subport */
-  for (subport_id = 0; subport_id < hqos->port.n_subports_per_port;
-       subport_id++)
-    {
-      u32 pipe_id;
-
-      rv =
-       rte_sched_subport_config (xd->hqos_ht->hqos, subport_id,
-                                 &hqos->subport[subport_id]);
-      if (rv)
-       return clib_error_return (0,
-                                 "HQoS%u subport %u: rte_sched_subport_config err (%d)",
-                                 xd->port_id, subport_id, rv);
-
-      /* HQoS pipe */
-      for (pipe_id = 0; pipe_id < hqos->port.n_pipes_per_subport; pipe_id++)
-       {
-         u32 pos = subport_id * hqos->port.n_pipes_per_subport + pipe_id;
-         u32 profile_id = hqos->pipe_map[pos];
-
-         rv =
-           rte_sched_pipe_config (xd->hqos_ht->hqos, subport_id, pipe_id,
-                                  profile_id);
-         if (rv)
-           return clib_error_return (0,
-                                     "HQoS%u subport %u pipe %u: rte_sched_pipe_config err (%d)",
-                                     xd->port_id, subport_id, pipe_id, rv);
-       }
-    }
-
-  /* Set up per-thread device data for the I/O TX thread */
-  xd->hqos_ht->hqos_burst_enq = hqos->burst_enq;
-  xd->hqos_ht->hqos_burst_deq = hqos->burst_deq;
-  vec_validate (xd->hqos_ht->pkts_enq, 2 * hqos->burst_enq - 1);
-  vec_validate (xd->hqos_ht->pkts_deq, hqos->burst_deq - 1);
-  xd->hqos_ht->pkts_enq_len = 0;
-  xd->hqos_ht->swq_pos = 0;
-  xd->hqos_ht->flush_count = 0;
-
-  /* Set up per-thread device data for each worker thread */
-  for (i = 0; i < worker_thread_count + 1; i++)
-    {
-      u32 tid;
-      if (i)
-       tid = worker_thread_first + (i - 1);
-      else
-       tid = i;
-
-      xd->hqos_wt[tid].swq = xd->hqos_ht->swq[i];
-      xd->hqos_wt[tid].hqos_field0_slabpos = hqos->pktfield0_slabpos;
-      xd->hqos_wt[tid].hqos_field0_slabmask = hqos->pktfield0_slabmask;
-      xd->hqos_wt[tid].hqos_field0_slabshr =
-       count_trailing_zeros (hqos->pktfield0_slabmask);
-      xd->hqos_wt[tid].hqos_field1_slabpos = hqos->pktfield1_slabpos;
-      xd->hqos_wt[tid].hqos_field1_slabmask = hqos->pktfield1_slabmask;
-      xd->hqos_wt[tid].hqos_field1_slabshr =
-       count_trailing_zeros (hqos->pktfield1_slabmask);
-      xd->hqos_wt[tid].hqos_field2_slabpos = hqos->pktfield2_slabpos;
-      xd->hqos_wt[tid].hqos_field2_slabmask = hqos->pktfield2_slabmask;
-      xd->hqos_wt[tid].hqos_field2_slabshr =
-       count_trailing_zeros (hqos->pktfield2_slabmask);
-      memcpy (xd->hqos_wt[tid].hqos_tc_table, hqos->tc_table,
-             sizeof (hqos->tc_table));
-    }
-
-  return 0;
-}
-
-/***
- *
- * HQoS run-time
- *
- ***/
-/*
- * dpdk_hqos_thread - Contains the main loop of an HQoS thread.
- *
- * w
- *     Information for the current thread
- */
-static_always_inline void
-dpdk_hqos_thread_internal_hqos_dbg_bypass (vlib_main_t * vm)
-{
-  dpdk_main_t *dm = &dpdk_main;
-  u32 thread_index = vm->thread_index;
-  u32 dev_pos;
-
-  dev_pos = 0;
-  while (1)
-    {
-      vlib_worker_thread_barrier_check ();
-
-      u32 n_devs = vec_len (dm->devices_by_hqos_cpu[thread_index]);
-      if (dev_pos >= n_devs)
-       dev_pos = 0;
-
-      dpdk_device_and_queue_t *dq =
-       vec_elt_at_index (dm->devices_by_hqos_cpu[thread_index], dev_pos);
-      dpdk_device_t *xd = vec_elt_at_index (dm->devices, dq->device);
-
-      dpdk_device_hqos_per_hqos_thread_t *hqos = xd->hqos_ht;
-      u32 device_index = xd->port_id;
-      u16 queue_id = dq->queue_id;
-
-      struct rte_mbuf **pkts_enq = hqos->pkts_enq;
-      u32 pkts_enq_len = hqos->pkts_enq_len;
-      u32 swq_pos = hqos->swq_pos;
-      u32 n_swq = vec_len (hqos->swq), i;
-      u32 flush_count = hqos->flush_count;
-
-      for (i = 0; i < n_swq; i++)
-       {
-         /* Get current SWQ for this device */
-         struct rte_ring *swq = hqos->swq[swq_pos];
-
-         /* Read SWQ burst to packet buffer of this device */
-         pkts_enq_len += rte_ring_sc_dequeue_burst (swq,
-                                                    (void **)
-                                                    &pkts_enq[pkts_enq_len],
-                                                    hqos->hqos_burst_enq, 0);
-
-         /* Get next SWQ for this device */
-         swq_pos++;
-         if (swq_pos >= n_swq)
-           swq_pos = 0;
-         hqos->swq_pos = swq_pos;
-
-         /* HWQ TX enqueue when burst available */
-         if (pkts_enq_len >= hqos->hqos_burst_enq)
-           {
-             u32 n_pkts = rte_eth_tx_burst (device_index,
-                                            (uint16_t) queue_id,
-                                            pkts_enq,
-                                            (uint16_t) pkts_enq_len);
-
-             for (; n_pkts < pkts_enq_len; n_pkts++)
-               rte_pktmbuf_free (pkts_enq[n_pkts]);
-
-             pkts_enq_len = 0;
-             flush_count = 0;
-             break;
-           }
-       }
-      if (pkts_enq_len)
-       {
-         flush_count++;
-         if (PREDICT_FALSE (flush_count == HQOS_FLUSH_COUNT_THRESHOLD))
-           {
-             rte_sched_port_enqueue (hqos->hqos, pkts_enq, pkts_enq_len);
-
-             pkts_enq_len = 0;
-             flush_count = 0;
-           }
-       }
-      hqos->pkts_enq_len = pkts_enq_len;
-      hqos->flush_count = flush_count;
-
-      /* Advance to next device */
-      dev_pos++;
-    }
-}
-
-static_always_inline void
-dpdk_hqos_thread_internal (vlib_main_t * vm)
-{
-  dpdk_main_t *dm = &dpdk_main;
-  u32 thread_index = vm->thread_index;
-  u32 dev_pos;
-
-  dev_pos = 0;
-  while (1)
-    {
-      vlib_worker_thread_barrier_check ();
-
-      u32 n_devs = vec_len (dm->devices_by_hqos_cpu[thread_index]);
-      if (PREDICT_FALSE (n_devs == 0))
-       {
-         dev_pos = 0;
-         continue;
-       }
-      if (dev_pos >= n_devs)
-       dev_pos = 0;
-
-      dpdk_device_and_queue_t *dq =
-       vec_elt_at_index (dm->devices_by_hqos_cpu[thread_index], dev_pos);
-      dpdk_device_t *xd = vec_elt_at_index (dm->devices, dq->device);
-
-      dpdk_device_hqos_per_hqos_thread_t *hqos = xd->hqos_ht;
-      u32 device_index = xd->port_id;
-      u16 queue_id = dq->queue_id;
-
-      struct rte_mbuf **pkts_enq = hqos->pkts_enq;
-      struct rte_mbuf **pkts_deq = hqos->pkts_deq;
-      u32 pkts_enq_len = hqos->pkts_enq_len;
-      u32 swq_pos = hqos->swq_pos;
-      u32 n_swq = vec_len (hqos->swq), i;
-      u32 flush_count = hqos->flush_count;
-
-      /*
-       * SWQ dequeue and HQoS enqueue for current device
-       */
-      for (i = 0; i < n_swq; i++)
-       {
-         /* Get current SWQ for this device */
-         struct rte_ring *swq = hqos->swq[swq_pos];
-
-         /* Read SWQ burst to packet buffer of this device */
-         pkts_enq_len += rte_ring_sc_dequeue_burst (swq,
-                                                    (void **)
-                                                    &pkts_enq[pkts_enq_len],
-                                                    hqos->hqos_burst_enq, 0);
-
-         /* Get next SWQ for this device */
-         swq_pos++;
-         if (swq_pos >= n_swq)
-           swq_pos = 0;
-         hqos->swq_pos = swq_pos;
-
-         /* HQoS enqueue when burst available */
-         if (pkts_enq_len >= hqos->hqos_burst_enq)
-           {
-             rte_sched_port_enqueue (hqos->hqos, pkts_enq, pkts_enq_len);
-
-             pkts_enq_len = 0;
-             flush_count = 0;
-             break;
-           }
-       }
-      if (pkts_enq_len)
-       {
-         flush_count++;
-         if (PREDICT_FALSE (flush_count == HQOS_FLUSH_COUNT_THRESHOLD))
-           {
-             rte_sched_port_enqueue (hqos->hqos, pkts_enq, pkts_enq_len);
-
-             pkts_enq_len = 0;
-             flush_count = 0;
-           }
-       }
-      hqos->pkts_enq_len = pkts_enq_len;
-      hqos->flush_count = flush_count;
-
-      /*
-       * HQoS dequeue and HWQ TX enqueue for current device
-       */
-      {
-       u32 pkts_deq_len, n_pkts;
-
-       pkts_deq_len = rte_sched_port_dequeue (hqos->hqos,
-                                              pkts_deq,
-                                              hqos->hqos_burst_deq);
-
-       for (n_pkts = 0; n_pkts < pkts_deq_len;)
-         n_pkts += rte_eth_tx_burst (device_index,
-                                     (uint16_t) queue_id,
-                                     &pkts_deq[n_pkts],
-                                     (uint16_t) (pkts_deq_len - n_pkts));
-      }
-
-      /* Advance to next device */
-      dev_pos++;
-    }
-}
-
-void
-dpdk_hqos_thread (vlib_worker_thread_t * w)
-{
-  vlib_main_t *vm;
-  vlib_thread_main_t *tm = vlib_get_thread_main ();
-  dpdk_main_t *dm = &dpdk_main;
-
-  vm = vlib_get_main ();
-
-  ASSERT (vm->thread_index == vlib_get_thread_index ());
-
-  clib_time_init (&vm->clib_time);
-  clib_mem_set_heap (w->thread_mheap);
-
-  /* Wait until the dpdk init sequence is complete */
-  while (tm->worker_thread_release == 0)
-    vlib_worker_thread_barrier_check ();
-
-  if (vec_len (dm->devices_by_hqos_cpu[vm->thread_index]) == 0)
-    return
-      clib_error
-      ("current I/O TX thread does not have any devices assigned to it");
-
-  if (DPDK_HQOS_DBG_BYPASS)
-    dpdk_hqos_thread_internal_hqos_dbg_bypass (vm);
-  else
-    dpdk_hqos_thread_internal (vm);
-}
-
-void
-dpdk_hqos_thread_fn (void *arg)
-{
-  vlib_worker_thread_t *w = (vlib_worker_thread_t *) arg;
-  vlib_worker_thread_init (w);
-  dpdk_hqos_thread (w);
-}
-
-/* *INDENT-OFF* */
-VLIB_REGISTER_THREAD (hqos_thread_reg, static) =
-{
-  .name = "hqos-threads",
-  .short_name = "hqos-threads",
-  .function = dpdk_hqos_thread_fn,
-};
-/* *INDENT-ON* */
-
-/*
- * HQoS run-time code to be called by the worker threads
- */
-#define BITFIELD(byte_array, slab_pos, slab_mask, slab_shr)     \
-({                                                              \
-  u64 slab = *((u64 *) &byte_array[slab_pos]);                  \
-  u64 val = (rte_be_to_cpu_64(slab) & slab_mask) >> slab_shr;   \
-  val;                                                          \
-})
-
-#define RTE_SCHED_PORT_HIERARCHY(subport, pipe, traffic_class, queue, color) \
-  ((((u64) (queue)) & 0x3) |                               \
-  ((((u64) (traffic_class)) & 0x3) << 2) |                 \
-  ((((u64) (color)) & 0x3) << 4) |                         \
-  ((((u64) (subport)) & 0xFFFF) << 16) |                   \
-  ((((u64) (pipe)) & 0xFFFFFFFF) << 32))
-
-void
-dpdk_hqos_metadata_set (dpdk_device_hqos_per_worker_thread_t * hqos,
-                       struct rte_mbuf **pkts, u32 n_pkts)
-{
-  u32 i;
-
-  for (i = 0; i < (n_pkts & (~0x3)); i += 4)
-    {
-      struct rte_mbuf *pkt0 = pkts[i];
-      struct rte_mbuf *pkt1 = pkts[i + 1];
-      struct rte_mbuf *pkt2 = pkts[i + 2];
-      struct rte_mbuf *pkt3 = pkts[i + 3];
-
-      u8 *pkt0_data = rte_pktmbuf_mtod (pkt0, u8 *);
-      u8 *pkt1_data = rte_pktmbuf_mtod (pkt1, u8 *);
-      u8 *pkt2_data = rte_pktmbuf_mtod (pkt2, u8 *);
-      u8 *pkt3_data = rte_pktmbuf_mtod (pkt3, u8 *);
-
-      u64 pkt0_subport = BITFIELD (pkt0_data, hqos->hqos_field0_slabpos,
-                                  hqos->hqos_field0_slabmask,
-                                  hqos->hqos_field0_slabshr);
-      u64 pkt0_pipe = BITFIELD (pkt0_data, hqos->hqos_field1_slabpos,
-                               hqos->hqos_field1_slabmask,
-                               hqos->hqos_field1_slabshr);
-      u64 pkt0_dscp = BITFIELD (pkt0_data, hqos->hqos_field2_slabpos,
-                               hqos->hqos_field2_slabmask,
-                               hqos->hqos_field2_slabshr);
-      u32 pkt0_tc = hqos->hqos_tc_table[pkt0_dscp & 0x3F] >> 2;
-      u32 pkt0_tc_q = hqos->hqos_tc_table[pkt0_dscp & 0x3F] & 0x3;
-
-      u64 pkt1_subport = BITFIELD (pkt1_data, hqos->hqos_field0_slabpos,
-                                  hqos->hqos_field0_slabmask,
-                                  hqos->hqos_field0_slabshr);
-      u64 pkt1_pipe = BITFIELD (pkt1_data, hqos->hqos_field1_slabpos,
-                               hqos->hqos_field1_slabmask,
-                               hqos->hqos_field1_slabshr);
-      u64 pkt1_dscp = BITFIELD (pkt1_data, hqos->hqos_field2_slabpos,
-                               hqos->hqos_field2_slabmask,
-                               hqos->hqos_field2_slabshr);
-      u32 pkt1_tc = hqos->hqos_tc_table[pkt1_dscp & 0x3F] >> 2;
-      u32 pkt1_tc_q = hqos->hqos_tc_table[pkt1_dscp & 0x3F] & 0x3;
-
-      u64 pkt2_subport = BITFIELD (pkt2_data, hqos->hqos_field0_slabpos,
-                                  hqos->hqos_field0_slabmask,
-                                  hqos->hqos_field0_slabshr);
-      u64 pkt2_pipe = BITFIELD (pkt2_data, hqos->hqos_field1_slabpos,
-                               hqos->hqos_field1_slabmask,
-                               hqos->hqos_field1_slabshr);
-      u64 pkt2_dscp = BITFIELD (pkt2_data, hqos->hqos_field2_slabpos,
-                               hqos->hqos_field2_slabmask,
-                               hqos->hqos_field2_slabshr);
-      u32 pkt2_tc = hqos->hqos_tc_table[pkt2_dscp & 0x3F] >> 2;
-      u32 pkt2_tc_q = hqos->hqos_tc_table[pkt2_dscp & 0x3F] & 0x3;
-
-      u64 pkt3_subport = BITFIELD (pkt3_data, hqos->hqos_field0_slabpos,
-                                  hqos->hqos_field0_slabmask,
-                                  hqos->hqos_field0_slabshr);
-      u64 pkt3_pipe = BITFIELD (pkt3_data, hqos->hqos_field1_slabpos,
-                               hqos->hqos_field1_slabmask,
-                               hqos->hqos_field1_slabshr);
-      u64 pkt3_dscp = BITFIELD (pkt3_data, hqos->hqos_field2_slabpos,
-                               hqos->hqos_field2_slabmask,
-                               hqos->hqos_field2_slabshr);
-      u32 pkt3_tc = hqos->hqos_tc_table[pkt3_dscp & 0x3F] >> 2;
-      u32 pkt3_tc_q = hqos->hqos_tc_table[pkt3_dscp & 0x3F] & 0x3;
-
-      u64 pkt0_sched = RTE_SCHED_PORT_HIERARCHY (pkt0_subport,
-                                                pkt0_pipe,
-                                                pkt0_tc,
-                                                pkt0_tc_q,
-                                                0);
-      u64 pkt1_sched = RTE_SCHED_PORT_HIERARCHY (pkt1_subport,
-                                                pkt1_pipe,
-                                                pkt1_tc,
-                                                pkt1_tc_q,
-                                                0);
-      u64 pkt2_sched = RTE_SCHED_PORT_HIERARCHY (pkt2_subport,
-                                                pkt2_pipe,
-                                                pkt2_tc,
-                                                pkt2_tc_q,
-                                                0);
-      u64 pkt3_sched = RTE_SCHED_PORT_HIERARCHY (pkt3_subport,
-                                                pkt3_pipe,
-                                                pkt3_tc,
-                                                pkt3_tc_q,
-                                                0);
-
-      pkt0->hash.sched.lo = pkt0_sched & 0xFFFFFFFF;
-      pkt0->hash.sched.hi = pkt0_sched >> 32;
-      pkt1->hash.sched.lo = pkt1_sched & 0xFFFFFFFF;
-      pkt1->hash.sched.hi = pkt1_sched >> 32;
-      pkt2->hash.sched.lo = pkt2_sched & 0xFFFFFFFF;
-      pkt2->hash.sched.hi = pkt2_sched >> 32;
-      pkt3->hash.sched.lo = pkt3_sched & 0xFFFFFFFF;
-      pkt3->hash.sched.hi = pkt3_sched >> 32;
-    }
-
-  for (; i < n_pkts; i++)
-    {
-      struct rte_mbuf *pkt = pkts[i];
-
-      u8 *pkt_data = rte_pktmbuf_mtod (pkt, u8 *);
-
-      u64 pkt_subport = BITFIELD (pkt_data, hqos->hqos_field0_slabpos,
-                                 hqos->hqos_field0_slabmask,
-                                 hqos->hqos_field0_slabshr);
-      u64 pkt_pipe = BITFIELD (pkt_data, hqos->hqos_field1_slabpos,
-                              hqos->hqos_field1_slabmask,
-                              hqos->hqos_field1_slabshr);
-      u64 pkt_dscp = BITFIELD (pkt_data, hqos->hqos_field2_slabpos,
-                              hqos->hqos_field2_slabmask,
-                              hqos->hqos_field2_slabshr);
-      u32 pkt_tc = hqos->hqos_tc_table[pkt_dscp & 0x3F] >> 2;
-      u32 pkt_tc_q = hqos->hqos_tc_table[pkt_dscp & 0x3F] & 0x3;
-
-      u64 pkt_sched = RTE_SCHED_PORT_HIERARCHY (pkt_subport,
-                                               pkt_pipe,
-                                               pkt_tc,
-                                               pkt_tc_q,
-                                               0);
-
-      pkt->hash.sched.lo = pkt_sched & 0xFFFFFFFF;
-      pkt->hash.sched.hi = pkt_sched >> 32;
-    }
-}
-
-/*
- * fd.io coding-style-patch-verification: ON
- *
- * Local Variables:
- * eval: (c-set-style "gnu")
- * End:
- */