dpdk_device_stop (xd);
}
+ /* Enable flow director when flows exist */
+ if (xd->pmd == VNET_DPDK_PMD_I40E)
+ {
+ if ((xd->flags & DPDK_DEVICE_FLAG_RX_FLOW_OFFLOAD) != 0)
+ xd->port_conf.fdir_conf.mode = RTE_FDIR_MODE_PERFECT;
+ else
+ xd->port_conf.fdir_conf.mode = RTE_FDIR_MODE_NONE;
+ }
+
rv = rte_eth_dev_configure (xd->device_index, xd->rx_q_used,
xd->tx_q_used, &xd->port_conf);
memset (fe, 0, sizeof (*fe));
pool_put (xd->flow_entries, fe);
- return 0;
+
+ goto disable_rx_offload;
}
if (op != VNET_FLOW_DEV_OP_ADD_FLOW)
else
fe->mark = 0;
+ if ((xd->flags & DPDK_DEVICE_FLAG_RX_FLOW_OFFLOAD) == 0)
+ {
+ xd->flags |= DPDK_DEVICE_FLAG_RX_FLOW_OFFLOAD;
+ dpdk_device_setup (xd);
+ }
+
switch (flow->type)
{
case VNET_FLOW_TYPE_IP4_N_TUPLE:
pool_put (xd->flow_lookup_entries, fle);
}
}
+disable_rx_offload:
+ if ((xd->flags & DPDK_DEVICE_FLAG_RX_FLOW_OFFLOAD) != 0
+ && pool_elts (xd->flow_entries) == 0)
+ {
+ xd->flags &= ~DPDK_DEVICE_FLAG_RX_FLOW_OFFLOAD;
+ dpdk_device_setup (xd);
+ }
+
return rv;
}
case VNET_DPDK_PMD_IXGBE:
case VNET_DPDK_PMD_I40E:
xd->port_type = port_type_from_speed_capa (&dev_info);
-#if 0
- xd->port_conf.fdir_conf.mode = RTE_FDIR_MODE_PERFECT;
-#endif
xd->supported_flow_actions = VNET_FLOW_ACTION_MARK |
VNET_FLOW_ACTION_REDIRECT_TO_NODE |
VNET_FLOW_ACTION_BUFFER_ADVANCE |