hw = vnet_get_hw_interface (dm->vnet_main, hw_if_index);
xd = vec_elt_at_index (dm->devices, hw->dev_instance);
- if (xd->dev_type != VNET_DPDK_DEV_ETH)
+ if ((xd->flags & DPDK_DEVICE_FLAG_PMD) == 0)
return clib_error_return (0, "number of descriptors can be set only for "
"physical devices");
* This device only supports one TX queue,
* and we're running multi-threaded...
*/
- if (PREDICT_FALSE (xd->dev_type != VNET_DPDK_DEV_VHOST_USER &&
+ if (PREDICT_FALSE ((xd->flags & DPDK_DEVICE_FLAG_VHOST_USER) == 0 &&
xd->lockp != 0))
{
queue_id = queue_id % xd->tx_q_used;
queue_id = (queue_id + 1) % xd->tx_q_used;
}
- if (PREDICT_TRUE (xd->dev_type == VNET_DPDK_DEV_ETH))
+ if (PREDICT_TRUE (xd->flags & DPDK_DEVICE_FLAG_PMD))
{
if (PREDICT_TRUE (tx_head > tx_tail))
{
}
}
#if DPDK_VHOST_USER
- else if (xd->dev_type == VNET_DPDK_DEV_VHOST_USER)
+ else if (xd->flags & DPDK_DEVICE_FLAG_VHOST_USER)
{
u32 offset = 0;
if (xd->need_txlock)
}
#endif
#if RTE_LIBRTE_KNI
- else if (xd->dev_type == VNET_DPDK_DEV_KNI)
+ else if (xd->flags & DPDK_DEVICE_FLAG_KNI)
{
if (PREDICT_TRUE (tx_head > tx_tail))
{
rv = 0;
}
- if (PREDICT_FALSE (xd->dev_type != VNET_DPDK_DEV_VHOST_USER &&
+ if (PREDICT_FALSE ((xd->flags & DPDK_DEVICE_FLAG_VHOST_USER) == 0 &&
xd->lockp != 0))
*xd->lockp[queue_id] = 0;
dpdk_main_t *dm = &dpdk_main;
dpdk_device_t *xd = vec_elt_at_index (dm->devices, hi->dev_instance);
- if (!xd || xd->dev_type != VNET_DPDK_DEV_VHOST_USER)
+ if (!xd || (xd->flags & DPDK_DEVICE_FLAG_VHOST_USER) == 0)
{
clib_warning
("cannot renumber non-vhost-user interface (sw_if_index: %d)",
sizeof (xd->last_cleared_xstats[0]));
#if DPDK_VHOST_USER
- if (PREDICT_FALSE (xd->dev_type == VNET_DPDK_DEV_VHOST_USER))
+ if (PREDICT_FALSE (xd->flags & DPDK_DEVICE_FLAG_VHOST_USER))
{
int i;
for (i = 0; i < xd->rx_q_used * VIRTIO_QNUM; i++)
int rv = 0;
#ifdef RTE_LIBRTE_KNI
- if (xd->dev_type == VNET_DPDK_DEV_KNI)
+ if (xd->flags & DPDK_DEVICE_FLAG_KNI)
{
if (is_up)
{
}
#endif
#if DPDK_VHOST_USER
- if (xd->dev_type == VNET_DPDK_DEV_VHOST_USER)
+ if (xd->flags & DPDK_DEVICE_FLAG_VHOST_USER)
{
if (is_up)
{
else if (xd->vlan_subifs)
xd->vlan_subifs--;
- if (xd->dev_type != VNET_DPDK_DEV_ETH)
+ if ((xd->flags & DPDK_DEVICE_FLAG_PMD) == 0)
return 0;
/* currently we program VLANS only for IXGBE VF and I40E VF */
extern vlib_node_registration_t dpdk_input_node;
extern vlib_node_registration_t handoff_dispatch_node;
-typedef enum
-{
- VNET_DPDK_DEV_ETH = 1, /* Standard DPDK PMD driver */
- VNET_DPDK_DEV_KNI, /* Kernel NIC Interface */
- VNET_DPDK_DEV_VHOST_USER,
- VNET_DPDK_DEV_UNKNOWN, /* must be last */
-} dpdk_device_type_t;
-
#define foreach_dpdk_pmd \
_ ("rte_nicvf_pmd", THUNDERX) \
_ ("rte_em_pmd", E1000EM) \
/* number of sub-interfaces */
u16 vlan_subifs;
- dpdk_device_type_t dev_type:8;
dpdk_pmd_t pmd:8;
i8 cpu_socket;
u16 flags;
-#define DPDK_DEVICE_FLAG_ADMIN_UP (1 << 0)
-#define DPDK_DEVICE_FLAG_PROMISC (1 << 1)
+#define DPDK_DEVICE_FLAG_ADMIN_UP (1 << 0)
+#define DPDK_DEVICE_FLAG_PROMISC (1 << 1)
+#define DPDK_DEVICE_FLAG_PMD (1 << 2)
+#define DPDK_DEVICE_FLAG_KNI (1 << 3)
+#define DPDK_DEVICE_FLAG_VHOST_USER (1 << 4)
CLIB_CACHE_LINE_ALIGN_MARK (cacheline1);
u64 rxerrors, last_rxerrors;
/* only update counters for PMD interfaces */
- if (xd->dev_type != VNET_DPDK_DEV_ETH)
+ if ((xd->flags & DPDK_DEVICE_FLAG_PMD) == 0)
return;
xd->time_last_stats_update = now ? now : xd->time_last_stats_update;
devname_format = "%s%x/%x/%x";
#ifdef RTE_LIBRTE_KNI
- if (dm->devices[i].dev_type == VNET_DPDK_DEV_KNI)
+ if (dm->devices[i].flags & DPDK_DEVICE_FLAG_KNI)
{
return format (s, "kni%d", dm->devices[i].kni_port_id);
}
else
#endif
#if DPDK_VHOST_USER
- if (dm->devices[i].dev_type == VNET_DPDK_DEV_VHOST_USER)
+ if (dm->devices[i].flags & DPDK_DEVICE_FLAG_VHOST_USER)
{
return format (s, "VirtualEthernet0/0/%d", dm->devices[i].vu_if_id);
}
char *dev_type;
u32 i = va_arg (*args, u32);
- if (dm->devices[i].dev_type == VNET_DPDK_DEV_KNI)
+ if (dm->devices[i].flags & DPDK_DEVICE_FLAG_KNI)
{
return format (s, "Kernel NIC Interface");
}
- else if (dm->devices[i].dev_type == VNET_DPDK_DEV_VHOST_USER)
+ else if (dm->devices[i].flags & DPDK_DEVICE_FLAG_VHOST_USER)
{
return format (s, "vhost-user interface");
}
rte_eth_dev_info_get (xd->device_index, &di);
- if (verbose > 1 && xd->dev_type == VNET_DPDK_DEV_ETH)
+ if (verbose > 1 && xd->flags & DPDK_DEVICE_FLAG_PMD)
{
struct rte_pci_device *pci;
struct rte_eth_rss_conf rss_conf;
format_dpdk_rss_hf_name, di.flow_type_rss_offloads);
}
- if (verbose && xd->dev_type == VNET_DPDK_DEV_VHOST_USER)
+ if (verbose && xd->flags & DPDK_DEVICE_FLAG_VHOST_USER)
{
s = format (s, "%Uqueue size (max): rx %d (%d) tx %d (%d)\n",
format_white_space, indent + 2,
#endif
#if DPDK_VHOST_USER
- if (verbose && xd->dev_type == VNET_DPDK_DEV_VHOST_USER)
+ if (verbose && xd->flags & DPDK_DEVICE_FLAG_VHOST_USER)
{
int i;
for (i = 0; i < xd->rx_q_used * VIRTIO_QNUM; i++)
else
xd->rx_q_used = 1;
- xd->dev_type = VNET_DPDK_DEV_ETH;
+ xd->flags |= DPDK_DEVICE_FLAG_PMD;
/* workaround for drivers not setting driver_name */
if ((!dev_info.driver_name) && (dev_info.pci_dev))
/* Create vnet interface */
vec_add2_aligned (dm->devices, xd, 1, CLIB_CACHE_LINE_BYTES);
- xd->dev_type = VNET_DPDK_DEV_KNI;
+ xd->flags |= DPDK_DEVICE_FLAG_KNI;
xd->device_index = xd - dm->devices;
ASSERT (nports + i == xd->device_index);
u8 hw_flags_chg = 0;
/* only update link state for PMD interfaces */
- if (xd->dev_type != VNET_DPDK_DEV_ETH)
+ if ((xd->flags & DPDK_DEVICE_FLAG_PMD) == 0)
return;
xd->time_last_link_update = now ? now : xd->time_last_link_update;
dpdk_update_link_state (xd, now);
#if DPDK_VHOST_USER
- if (xd->dev_type == VNET_DPDK_DEV_VHOST_USER)
+ if (xd->flags & DPDK_DEVICE_FLAG_VHOST_USER)
if (dpdk_vhost_user_process_if (vm, xd, vu_state) != 0)
continue;
#endif
n_left = VLIB_FRAME_SIZE;
n_buffers = 0;
- if (PREDICT_TRUE (xd->dev_type == VNET_DPDK_DEV_ETH))
+ if (PREDICT_TRUE (xd->flags & DPDK_DEVICE_FLAG_PMD))
{
while (n_left)
{
}
}
#if DPDK_VHOST_USER
- else if (xd->dev_type == VNET_DPDK_DEV_VHOST_USER)
+ else if (xd->flags & DPDK_DEVICE_FLAG_VHOST_USER)
{
vlib_main_t *vm = vlib_get_main ();
vlib_buffer_main_t *bm = vm->buffer_main;
}
#endif
#ifdef RTE_LIBRTE_KNI
- else if (xd->dev_type == VNET_DPDK_DEV_KNI)
+ else if (xd->flags & DPDK_DEVICE_FLAG_KNI)
{
n_buffers =
rte_kni_rx_burst (xd->kni, xd->rx_vectors[queue_id], VLIB_FRAME_SIZE);
vnet_hw_interface_t *hi = vnet_get_hw_interface (vnm, hw_if_index);
dpdk_device_t *xd = vec_elt_at_index (dm->devices, hi->dev_instance);
- if (xd->dev_type != VNET_DPDK_DEV_VHOST_USER)
+ if ((xd->flags DPDK_DEVICE_FLAG_VHOST_USER) == 0)
return 0;
return xd;
if (vec_len (dm->devices) > vui_idx)
{
xd = vec_elt_at_index (dm->devices, vui_idx);
- if (xd->dev_type == VNET_DPDK_DEV_VHOST_USER)
+ if (xd->flags & DPDK_DEVICE_FLAG_VHOST_USER)
{
DBG_SOCK
("reusing inactive vhost-user interface sw_if_index %d",
{
// vui was not retrieved from inactive ifaces - create new
vec_add2_aligned (dm->devices, xd, 1, CLIB_CACHE_LINE_BYTES);
- xd->dev_type = VNET_DPDK_DEV_VHOST_USER;
+ xd->flags |= DPDK_DEVICE_FLAG_VHOST_USER;
xd->rx_q_used = num_qpairs;
xd->tx_q_used = num_qpairs;
xd->vu_vhost_dev.virt_qp_nb = num_qpairs;
vec_foreach (xd, dm->devices)
{
- if (xd->dev_type == VNET_DPDK_DEV_VHOST_USER && xd->vu_intf->active)
+ if ((xd->flags & DPDK_DEVICE_FLAG_VHOST_USER) && xd->vu_intf->active)
vec_add1 (hw_if_indices, xd->vlib_hw_if_index);
}
{
vec_foreach (xd, dm->devices)
{
- if (xd->dev_type == VNET_DPDK_DEV_VHOST_USER && xd->vu_intf->active)
+ if ((xd->flags DPDK_DEVICE_FLAG_VHOST_USER) && xd->vu_intf->active)
vec_add1 (hw_if_indices, xd->vlib_hw_if_index);
}
}