dpdk_device_error (dpdk_device_t * xd, char *str, int rv)
{
dpdk_log_err ("Interface %U error %d: %s",
- format_dpdk_device_name, xd->device_index, rv,
- rte_strerror (rv));
+ format_dpdk_device_name, xd->port_id, rv, rte_strerror (rv));
xd->errors = clib_error_return (xd->errors, "%s[port:%d, errno:%d]: %s",
- str, xd->device_index, rv,
- rte_strerror (rv));
+ str, xd->port_id, rv, rte_strerror (rv));
}
void
vnet_main_t *vnm = vnet_get_main ();
vnet_sw_interface_t *sw = vnet_get_sw_interface (vnm, xd->sw_if_index);
vnet_hw_interface_t *hi = vnet_get_hw_interface (vnm, xd->hw_if_index);
+ struct rte_eth_dev_info dev_info;
+ u64 bitmap;
int rv;
int j;
xd->port_conf.fdir_conf.mode = RTE_FDIR_MODE_NONE;
}
- rv = rte_eth_dev_configure (xd->device_index, xd->rx_q_used,
+ rte_eth_dev_info_get (xd->port_id, &dev_info);
+
+ bitmap = xd->port_conf.txmode.offloads & ~dev_info.tx_offload_capa;
+ if (bitmap)
+ {
+ dpdk_log_warn ("unsupported tx offloads requested on port %u: %U",
+ xd->port_id, format_dpdk_tx_offload_caps, bitmap);
+ xd->port_conf.txmode.offloads ^= bitmap;
+ }
+
+ bitmap = xd->port_conf.rxmode.offloads & ~dev_info.rx_offload_capa;
+ if (bitmap)
+ {
+ dpdk_log_warn ("unsupported rx offloads requested on port %u: %U",
+ xd->port_id, format_dpdk_rx_offload_caps, bitmap);
+ xd->port_conf.rxmode.offloads ^= bitmap;
+ }
+
+ rv = rte_eth_dev_configure (xd->port_id, xd->rx_q_used,
xd->tx_q_used, &xd->port_conf);
if (rv < 0)
/* Set up one TX-queue per worker thread */
for (j = 0; j < xd->tx_q_used; j++)
{
- rv = rte_eth_tx_queue_setup (xd->device_index, j, xd->nb_tx_desc,
- xd->cpu_socket, &xd->tx_conf);
+ rv =
+ rte_eth_tx_queue_setup (xd->port_id, j, xd->nb_tx_desc,
+ xd->cpu_socket, &xd->tx_conf);
/* retry with any other CPU socket */
if (rv < 0)
- rv = rte_eth_tx_queue_setup (xd->device_index, j, xd->nb_tx_desc,
- SOCKET_ID_ANY, &xd->tx_conf);
+ rv =
+ rte_eth_tx_queue_setup (xd->port_id, j,
+ xd->nb_tx_desc, SOCKET_ID_ANY,
+ &xd->tx_conf);
if (rv < 0)
dpdk_device_error (xd, "rte_eth_tx_queue_setup", rv);
}
dpdk_mempool_private_t *privp;
uword tidx = vnet_get_device_input_thread_index (dm->vnet_main,
xd->hw_if_index, j);
- unsigned lcore = vlib_worker_threads[tidx].lcore_id;
+ unsigned lcore = vlib_worker_threads[tidx].cpu_id;
u16 socket_id = rte_lcore_to_socket_id (lcore);
- rv = rte_eth_rx_queue_setup (xd->device_index, j, xd->nb_rx_desc,
- xd->cpu_socket, 0,
- dm->pktmbuf_pools[socket_id]);
+ rv =
+ rte_eth_rx_queue_setup (xd->port_id, j, xd->nb_rx_desc,
+ xd->cpu_socket, 0,
+ dm->pktmbuf_pools[socket_id]);
/* retry with any other CPU socket */
if (rv < 0)
- rv = rte_eth_rx_queue_setup (xd->device_index, j, xd->nb_rx_desc,
- SOCKET_ID_ANY, 0,
- dm->pktmbuf_pools[socket_id]);
+ rv =
+ rte_eth_rx_queue_setup (xd->port_id, j,
+ xd->nb_rx_desc, SOCKET_ID_ANY, 0,
+ dm->pktmbuf_pools[socket_id]);
privp = rte_mempool_get_priv (dm->pktmbuf_pools[socket_id]);
xd->buffer_pool_for_queue[j] = privp->buffer_pool_index;
if (vec_len (xd->errors))
goto error;
- rte_eth_dev_set_mtu (xd->device_index, hi->max_packet_bytes);
+ rte_eth_dev_set_mtu (xd->port_id, hi->max_packet_bytes);
if (xd->flags & DPDK_DEVICE_FLAG_ADMIN_UP)
dpdk_device_start (xd);
if (xd->flags & DPDK_DEVICE_FLAG_PMD_INIT_FAIL)
return;
- rv = rte_eth_dev_start (xd->device_index);
+ rv = rte_eth_dev_start (xd->port_id);
if (rv)
{
if (xd->default_mac_address)
rv =
- rte_eth_dev_default_mac_addr_set (xd->device_index,
+ rte_eth_dev_default_mac_addr_set (xd->port_id,
(struct ether_addr *)
xd->default_mac_address);
dpdk_device_error (xd, "rte_eth_dev_default_mac_addr_set", rv);
if (xd->flags & DPDK_DEVICE_FLAG_PROMISC)
- rte_eth_promiscuous_enable (xd->device_index);
+ rte_eth_promiscuous_enable (xd->port_id);
else
- rte_eth_promiscuous_disable (xd->device_index);
+ rte_eth_promiscuous_disable (xd->port_id);
- rte_eth_allmulticast_enable (xd->device_index);
+ rte_eth_allmulticast_enable (xd->port_id);
if (xd->pmd == VNET_DPDK_PMD_BOND)
{
dpdk_portid_t slink[16];
- int nlink = rte_eth_bond_slaves_get (xd->device_index, slink, 16);
+ int nlink = rte_eth_bond_slaves_get (xd->port_id, slink, 16);
while (nlink >= 1)
{
dpdk_portid_t dpdk_port = slink[--nlink];
}
dpdk_log_info ("Interface %U started",
- format_dpdk_device_name, xd->device_index);
+ format_dpdk_device_name, xd->port_id);
}
void
if (xd->flags & DPDK_DEVICE_FLAG_PMD_INIT_FAIL)
return;
- rte_eth_allmulticast_disable (xd->device_index);
- rte_eth_dev_stop (xd->device_index);
+ rte_eth_allmulticast_disable (xd->port_id);
+ rte_eth_dev_stop (xd->port_id);
+ memset (&xd->link, 0, sizeof (struct rte_eth_link));
/* For bonded interface, stop slave links */
if (xd->pmd == VNET_DPDK_PMD_BOND)
{
dpdk_portid_t slink[16];
- int nlink = rte_eth_bond_slaves_get (xd->device_index, slink, 16);
+ int nlink = rte_eth_bond_slaves_get (xd->port_id, slink, 16);
while (nlink >= 1)
{
dpdk_portid_t dpdk_port = slink[--nlink];
}
}
dpdk_log_info ("Interface %U stopped",
- format_dpdk_device_name, xd->device_index);
+ format_dpdk_device_name, xd->port_id);
}
/* Even type for send_garp_na_process */
send_garp_na_process (vlib_main_t * vm,
vlib_node_runtime_t * rt, vlib_frame_t * f)
{
- vnet_main_t *vnm = vnet_get_main ();
uword event_type, *event_data = 0;
while (1)
if (i < 5) /* wait 0.2 sec for link to settle, max total 1 sec */
vlib_process_suspend (vm, 0.2);
dpdk_device_t *xd = &dpdk_main.devices[dpdk_port];
- u32 hw_if_index = xd->hw_if_index;
- vnet_hw_interface_t *hi = vnet_get_hw_interface (vnm, hw_if_index);
dpdk_update_link_state (xd, vlib_time_now (vm));
- send_ip4_garp (vm, hi);
- send_ip6_na (vm, hi);
+ send_ip4_garp (vm, xd->sw_if_index);
+ send_ip6_na (vm, xd->sw_if_index);
}
vec_reset_length (event_data);
}
dpdk_log_info ("Port %d state to %s, "
"slave of port %d BondEthernet%d in mode %d",
port_id, (link_up) ? "UP" : "DOWN",
- bd_port, xd->port_id, bd_mode);
+ bd_port, xd->bond_instance_num, bd_mode);
if (bd_mode == BONDING_MODE_ACTIVE_BACKUP)
{
vl_api_force_rpc_call_main_thread