/* $$$$ We should rename always_inline -> clib_always_inline */
#undef always_inline
+#define ALLOW_EXPERIMENTAL_API
+
#include <rte_config.h>
#include <rte_common.h>
#include <rte_mempool.h>
#include <rte_mbuf.h>
#include <rte_version.h>
-#include <rte_eth_bond.h>
#include <rte_sched.h>
+#include <rte_net.h>
+#include <rte_bus_pci.h>
+#include <rte_flow.h>
-#include <vnet/unix/pcap.h>
#include <vnet/devices/devices.h>
#if CLIB_DEBUG > 0
#endif
#include <vlib/pci/pci.h>
-
-#define NB_MBUF (16<<10)
+#include <vnet/flow/flow.h>
extern vnet_device_class_t dpdk_device_class;
extern vlib_node_registration_t dpdk_input_node;
+extern vlib_node_registration_t admin_up_down_process_node;
-#if RTE_VERSION >= RTE_VERSION_NUM(17, 2, 0, 0)
-#define foreach_dpdk_pmd \
- _ ("net_thunderx", THUNDERX) \
- _ ("net_e1000_em", E1000EM) \
- _ ("net_e1000_igb", IGB) \
- _ ("net_e1000_igb_vf", IGBVF) \
- _ ("net_ixgbe", IXGBE) \
- _ ("net_ixgbe_vf", IXGBEVF) \
- _ ("net_i40e", I40E) \
- _ ("net_i40e_vf", I40EVF) \
- _ ("net_virtio", VIRTIO) \
- _ ("net_enic", ENIC) \
- _ ("net_vmxnet3", VMXNET3) \
- _ ("AF_PACKET PMD", AF_PACKET) \
- _ ("net_bonding", BOND) \
- _ ("net_fm10k", FM10K) \
- _ ("net_cxgbe", CXGBE) \
- _ ("net_mlx5", MLX5) \
- _ ("net_dpaa2", DPAA2)
-#else
#define foreach_dpdk_pmd \
_ ("net_thunderx", THUNDERX) \
_ ("net_e1000_em", E1000EM) \
_ ("net_ixgbe_vf", IXGBEVF) \
_ ("net_i40e", I40E) \
_ ("net_i40e_vf", I40EVF) \
+ _ ("net_ice", ICE) \
+ _ ("net_iavf", IAVF) \
_ ("net_virtio", VIRTIO) \
_ ("net_enic", ENIC) \
_ ("net_vmxnet3", VMXNET3) \
_ ("AF_PACKET PMD", AF_PACKET) \
- _ ("rte_bond_pmd", BOND) \
_ ("net_fm10k", FM10K) \
_ ("net_cxgbe", CXGBE) \
+ _ ("net_mlx4", MLX4) \
_ ("net_mlx5", MLX5) \
- _ ("net_dpaa2", DPAA2)
-#endif
+ _ ("net_dpaa2", DPAA2) \
+ _ ("net_virtio_user", VIRTIO_USER) \
+ _ ("net_vhost", VHOST_ETHER) \
+ _ ("net_ena", ENA) \
+ _ ("net_failsafe", FAILSAFE) \
+ _ ("net_liovf", LIOVF_ETHER) \
+ _ ("net_qede", QEDE) \
+ _ ("net_netvsc", NETVSC) \
+ _ ("net_bnxt", BNXT)
typedef enum
{
typedef enum
{
VNET_DPDK_PORT_TYPE_ETH_1G,
+ VNET_DPDK_PORT_TYPE_ETH_2_5G,
+ VNET_DPDK_PORT_TYPE_ETH_5G,
VNET_DPDK_PORT_TYPE_ETH_10G,
+ VNET_DPDK_PORT_TYPE_ETH_20G,
VNET_DPDK_PORT_TYPE_ETH_25G,
VNET_DPDK_PORT_TYPE_ETH_40G,
+ VNET_DPDK_PORT_TYPE_ETH_50G,
+ VNET_DPDK_PORT_TYPE_ETH_56G,
VNET_DPDK_PORT_TYPE_ETH_100G,
- VNET_DPDK_PORT_TYPE_ETH_BOND,
VNET_DPDK_PORT_TYPE_ETH_SWITCH,
VNET_DPDK_PORT_TYPE_AF_PACKET,
+ VNET_DPDK_PORT_TYPE_ETH_VF,
+ VNET_DPDK_PORT_TYPE_VIRTIO_USER,
+ VNET_DPDK_PORT_TYPE_VHOST_ETHER,
+ VNET_DPDK_PORT_TYPE_FAILSAFE,
+ VNET_DPDK_PORT_TYPE_NETVSC,
VNET_DPDK_PORT_TYPE_UNKNOWN,
} dpdk_port_type_t;
-/*
- * The header for the tx_vector in dpdk_device_t.
- * Head and tail are indexes into the tx_vector and are of type
- * u64 so they never overflow.
- */
+typedef uint16_t dpdk_portid_t;
+
+#define foreach_dpdk_device_flags \
+ _( 0, ADMIN_UP, "admin-up") \
+ _( 1, PROMISC, "promisc") \
+ _( 2, PMD, "pmd") \
+ _( 3, PMD_INIT_FAIL, "pmd-init-fail") \
+ _( 4, MAYBE_MULTISEG, "maybe-multiseg") \
+ _( 5, HAVE_SUBIF, "subif") \
+ _( 9, TX_OFFLOAD, "tx-offload") \
+ _(10, INTEL_PHDR_CKSUM, "intel-phdr-cksum") \
+ _(11, RX_FLOW_OFFLOAD, "rx-flow-offload") \
+ _(12, RX_IP4_CKSUM, "rx-ip4-cksum")
+
+enum
+{
+#define _(a, b, c) DPDK_DEVICE_FLAG_##b = (1 << a),
+ foreach_dpdk_device_flags
+#undef _
+};
+
typedef struct
{
- u64 tx_head;
- u64 tx_tail;
-} tx_ring_hdr_t;
+ u32 flow_index;
+ u32 mark;
+ struct rte_flow *handle;
+} dpdk_flow_entry_t;
typedef struct
{
- struct rte_ring *swq;
-
- u64 hqos_field0_slabmask;
- u32 hqos_field0_slabpos;
- u32 hqos_field0_slabshr;
- u64 hqos_field1_slabmask;
- u32 hqos_field1_slabpos;
- u32 hqos_field1_slabshr;
- u64 hqos_field2_slabmask;
- u32 hqos_field2_slabpos;
- u32 hqos_field2_slabshr;
- u32 hqos_tc_table[64];
-} dpdk_device_hqos_per_worker_thread_t;
+ u32 flow_id;
+ u16 next_index;
+ i16 buffer_advance;
+} dpdk_flow_lookup_entry_t;
typedef struct
{
- struct rte_ring **swq;
- struct rte_mbuf **pkts_enq;
- struct rte_mbuf **pkts_deq;
- struct rte_sched_port *hqos;
- u32 hqos_burst_enq;
- u32 hqos_burst_deq;
- u32 pkts_enq_len;
- u32 swq_pos;
- u32 flush_count;
-} dpdk_device_hqos_per_hqos_thread_t;
+ CLIB_CACHE_LINE_ALIGN_MARK (cacheline0);
+ u8 buffer_pool_index;
+} dpdk_rx_queue_t;
typedef struct
{
CLIB_CACHE_LINE_ALIGN_MARK (cacheline0);
- volatile u32 **lockp;
+ clib_spinlock_t lock;
+} dpdk_tx_queue_t;
- /* Instance ID */
+typedef struct
+{
+ CLIB_CACHE_LINE_ALIGN_MARK (cacheline0);
+
+ dpdk_rx_queue_t *rx_queues;
+ dpdk_tx_queue_t *tx_queues;
+
+ /* Instance ID to access internal device array. */
u32 device_index;
u32 hw_if_index;
- u32 vlib_sw_if_index;
+ u32 sw_if_index;
/* next node index if we decide to steal the rx graph arc */
u32 per_interface_next_index;
- /* dpdk rte_mbuf rx and tx vectors, VLIB_FRAME_SIZE */
- struct rte_mbuf ***tx_vectors; /* one per worker thread */
- struct rte_mbuf ***rx_vectors;
-
- /* vector of traced contexts, per device */
- u32 **d_trace_buffers;
+ u16 rx_q_used;
+ u16 tx_q_used;
+ u16 flags;
+ /* DPDK device port number */
+ dpdk_portid_t port_id;
dpdk_pmd_t pmd:8;
i8 cpu_socket;
- u16 flags;
-#define DPDK_DEVICE_FLAG_ADMIN_UP (1 << 0)
-#define DPDK_DEVICE_FLAG_PROMISC (1 << 1)
-#define DPDK_DEVICE_FLAG_PMD (1 << 2)
-#define DPDK_DEVICE_FLAG_PMD_SUPPORTS_PTYPE (1 << 3)
-#define DPDK_DEVICE_FLAG_MAYBE_MULTISEG (1 << 4)
-#define DPDK_DEVICE_FLAG_HAVE_SUBIF (1 << 5)
-#define DPDK_DEVICE_FLAG_HQOS (1 << 6)
-
- u16 nb_tx_desc;
CLIB_CACHE_LINE_ALIGN_MARK (cacheline1);
+ u16 nb_tx_desc;
+ u16 nb_rx_desc;
+ u8 *name;
u8 *interface_name_suffix;
/* number of sub-interfaces */
u16 num_subifs;
/* PMD related */
- u16 tx_q_used;
- u16 rx_q_used;
- u16 nb_rx_desc;
- u16 *cpu_socket_id_by_queue;
struct rte_eth_conf port_conf;
struct rte_eth_txconf tx_conf;
- /* HQoS related */
- dpdk_device_hqos_per_worker_thread_t *hqos_wt;
- dpdk_device_hqos_per_hqos_thread_t *hqos_ht;
+ /* flow related */
+ u32 supported_flow_actions;
+ dpdk_flow_entry_t *flow_entries; /* pool */
+ dpdk_flow_lookup_entry_t *flow_lookup_entries; /* pool */
+ u32 *parked_lookup_indexes; /* vector */
+ u32 parked_loop_count;
+ struct rte_flow_error last_flow_error;
- /* af_packet or BondEthernet instance number */
- u8 port_id;
+ /* af_packet instance number */
+ u16 af_packet_instance_num;
struct rte_eth_link link;
f64 time_last_link_update;
struct rte_eth_stats stats;
struct rte_eth_stats last_stats;
- struct rte_eth_stats last_cleared_stats;
struct rte_eth_xstat *xstats;
- struct rte_eth_xstat *last_cleared_xstats;
f64 time_last_stats_update;
dpdk_port_type_t port_type;
/* mac address */
u8 *default_mac_address;
+
+ /* error string */
+ clib_error_t *errors;
} dpdk_device_t;
#define DPDK_STATS_POLL_INTERVAL (10.0)
#define DPDK_LINK_POLL_INTERVAL (3.0)
#define DPDK_MIN_LINK_POLL_INTERVAL (0.001) /* 1msec */
-typedef struct
-{
- u32 device;
- u16 queue_id;
-} dpdk_device_and_queue_t;
-
-#ifndef DPDK_HQOS_DBG_BYPASS
-#define DPDK_HQOS_DBG_BYPASS 0
-#endif
-
-#ifndef HQOS_FLUSH_COUNT_THRESHOLD
-#define HQOS_FLUSH_COUNT_THRESHOLD 100000
-#endif
-
-typedef struct dpdk_device_config_hqos_t
-{
- u32 hqos_thread;
- u32 hqos_thread_valid;
-
- u32 swq_size;
- u32 burst_enq;
- u32 burst_deq;
-
- u32 pktfield0_slabpos;
- u32 pktfield1_slabpos;
- u32 pktfield2_slabpos;
- u64 pktfield0_slabmask;
- u64 pktfield1_slabmask;
- u64 pktfield2_slabmask;
- u32 tc_table[64];
-
- struct rte_sched_port_params port;
- struct rte_sched_subport_params *subport;
- struct rte_sched_pipe_params *pipe;
- uint32_t *pipe_map;
-} dpdk_device_config_hqos_t;
-
-int dpdk_hqos_validate_mask (u64 mask, u32 n);
-void dpdk_device_config_hqos_pipe_profile_default (dpdk_device_config_hqos_t *
- hqos, u32 pipe_profile_id);
-void dpdk_device_config_hqos_default (dpdk_device_config_hqos_t * hqos);
-clib_error_t *dpdk_port_setup_hqos (dpdk_device_t * xd,
- dpdk_device_config_hqos_t * hqos);
-void dpdk_hqos_metadata_set (dpdk_device_hqos_per_worker_thread_t * hqos,
- struct rte_mbuf **pkts, u32 n_pkts);
-
#define foreach_dpdk_device_config_item \
_ (num_rx_queues) \
_ (num_tx_queues) \
typedef struct
{
vlib_pci_addr_t pci_addr;
+ u8 *name;
u8 is_blacklisted;
u8 vlan_strip_offload;
#define DPDK_DEVICE_VLAN_STRIP_DEFAULT 0
foreach_dpdk_device_config_item
#undef _
clib_bitmap_t * workers;
- u32 hqos_enabled;
- dpdk_device_config_hqos_t hqos;
+ u8 tso;
+ u8 *devargs;
+ clib_bitmap_t *rss_queues;
+
+#define DPDK_DEVICE_TSO_DEFAULT 0
+#define DPDK_DEVICE_TSO_OFF 1
+#define DPDK_DEVICE_TSO_ON 2
} dpdk_device_config_t;
typedef struct
u8 *uio_driver_name;
u8 no_multi_seg;
u8 enable_tcp_udp_checksum;
+ u8 no_tx_checksum_offload;
/* Required config parameters */
u8 coremask_set_manually;
u8 nchannels_set_manually;
u32 coremask;
u32 nchannels;
- u32 num_mbufs;
+ u32 num_crypto_mbufs;
/*
* format interface names ala xxxEthernet%d/%d/%d instead of
dpdk_device_config_t *dev_confs;
uword *device_config_index_by_pci_addr;
+ /* devices blacklist by pci vendor_id, device_id */
+ u32 *blacklist_by_pci_vendor_and_device;
+
} dpdk_config_main_t;
-dpdk_config_main_t dpdk_config_main;
+extern dpdk_config_main_t dpdk_config_main;
+
+#define DPDK_RX_BURST_SZ VLIB_FRAME_SIZE
+
+typedef struct
+{
+ CLIB_CACHE_LINE_ALIGN_MARK (cacheline0);
+ struct rte_mbuf *mbufs[DPDK_RX_BURST_SZ];
+ u32 buffers[DPDK_RX_BURST_SZ];
+ u16 next[DPDK_RX_BURST_SZ];
+ u16 etype[DPDK_RX_BURST_SZ];
+ u16 flags[DPDK_RX_BURST_SZ];
+ vlib_buffer_t buffer_template;
+} dpdk_per_thread_data_t;
typedef struct
{
/* Devices */
dpdk_device_t *devices;
- dpdk_device_and_queue_t **devices_by_hqos_cpu;
-
- /* per-thread recycle lists */
- u32 **recycle;
-
- /* per-thread buffer templates */
- vlib_buffer_t *buffer_templates;
+ dpdk_per_thread_data_t *per_thread_data;
/* buffer flags template, configurable to enable/disable tcp / udp cksum */
u32 buffer_flags_template;
- /* vlib buffer free list, must be same size as an rte_mbuf */
- u32 vlib_buffer_free_list_index;
-
- /* Ethernet input node index */
- u32 ethernet_input_node_index;
-
- /* pcap tracing [only works if (CLIB_DEBUG > 0)] */
- int tx_pcap_enable;
- pcap_main_t pcap_main;
- u8 *pcap_filename;
- u32 pcap_sw_if_index;
- u32 pcap_pkts_to_capture;
-
/*
* flag indicating that a posted admin up/down
* (via post_sw_interface_set_flags) is in progress
*/
u8 admin_up_down_in_progress;
- u8 use_rss;
-
- /* which cpus are running I/O TX */
- int hqos_cpu_first_index;
- int hqos_cpu_count;
-
/* control interval of dpdk link state and stat polling */
f64 link_state_poll_interval;
f64 stat_poll_interval;
- /* Sleep for this many usec after each device poll */
- u32 poll_sleep_usec;
-
/* convenience */
vlib_main_t *vlib_main;
vnet_main_t *vnet_main;
dpdk_config_main_t *conf;
- /* mempool */
- struct rte_mempool **pktmbuf_pools;
-
/* API message ID base */
u16 msg_id_base;
+
+ /* logging */
+ vlib_log_class_t log_default;
} dpdk_main_t;
extern dpdk_main_t dpdk_main;
struct rte_mbuf mb;
/* Copy of VLIB buffer; packet data stored in pre_data. */
vlib_buffer_t buffer;
-} dpdk_tx_dma_trace_t;
+ u8 data[256]; /* First 256 data bytes, used for hexdump */
+} dpdk_tx_trace_t;
typedef struct
{
struct rte_mbuf mb;
vlib_buffer_t buffer; /* Copy of VLIB buffer; pkt data stored in pre_data. */
u8 data[256]; /* First 256 data bytes, used for hexdump */
-} dpdk_rx_dma_trace_t;
+} dpdk_rx_trace_t;
-clib_error_t *dpdk_port_setup (dpdk_main_t * dm, dpdk_device_t * xd);
+void dpdk_device_setup (dpdk_device_t * xd);
+void dpdk_device_start (dpdk_device_t * xd);
+void dpdk_device_stop (dpdk_device_t * xd);
+int dpdk_port_state_callback (dpdk_portid_t port_id,
+ enum rte_eth_event_type type,
+ void *param, void *ret_param);
#define foreach_dpdk_error \
_(NONE, "no error") \
DPDK_N_ERROR,
} dpdk_error_t;
+#define dpdk_log_err(...) \
+ vlib_log(VLIB_LOG_LEVEL_ERR, dpdk_main.log_default, __VA_ARGS__)
+#define dpdk_log_warn(...) \
+ vlib_log(VLIB_LOG_LEVEL_WARNING, dpdk_main.log_default, __VA_ARGS__)
+#define dpdk_log_notice(...) \
+ vlib_log(VLIB_LOG_LEVEL_NOTICE, dpdk_main.log_default, __VA_ARGS__)
+#define dpdk_log_info(...) \
+ vlib_log(VLIB_LOG_LEVEL_INFO, dpdk_main.log_default, __VA_ARGS__)
+
void dpdk_update_link_state (dpdk_device_t * xd, f64 now);
+#define foreach_dpdk_rss_hf \
+ _(0, ETH_RSS_FRAG_IPV4, "ipv4-frag") \
+ _(1, ETH_RSS_NONFRAG_IPV4_TCP, "ipv4-tcp") \
+ _(2, ETH_RSS_NONFRAG_IPV4_UDP, "ipv4-udp") \
+ _(3, ETH_RSS_NONFRAG_IPV4_SCTP, "ipv4-sctp") \
+ _(4, ETH_RSS_NONFRAG_IPV4_OTHER, "ipv4-other") \
+ _(5, ETH_RSS_IPV4, "ipv4") \
+ _(6, ETH_RSS_IPV6_TCP_EX, "ipv6-tcp-ex") \
+ _(7, ETH_RSS_IPV6_UDP_EX, "ipv6-udp-ex") \
+ _(8, ETH_RSS_FRAG_IPV6, "ipv6-frag") \
+ _(9, ETH_RSS_NONFRAG_IPV6_TCP, "ipv6-tcp") \
+ _(10, ETH_RSS_NONFRAG_IPV6_UDP, "ipv6-udp") \
+ _(11, ETH_RSS_NONFRAG_IPV6_SCTP, "ipv6-sctp") \
+ _(12, ETH_RSS_NONFRAG_IPV6_OTHER, "ipv6-other") \
+ _(13, ETH_RSS_IPV6_EX, "ipv6-ex") \
+ _(14, ETH_RSS_IPV6, "ipv6") \
+ _(15, ETH_RSS_L2_PAYLOAD, "l2-payload") \
+ _(16, ETH_RSS_PORT, "port") \
+ _(17, ETH_RSS_VXLAN, "vxlan") \
+ _(18, ETH_RSS_GENEVE, "geneve") \
+ _(19, ETH_RSS_NVGRE, "nvgre") \
+ _(20, ETH_RSS_GTPU, "gtpu") \
+ _(60, ETH_RSS_L4_DST_ONLY, "l4-dst-only") \
+ _(61, ETH_RSS_L4_SRC_ONLY, "l4-src-only") \
+ _(62, ETH_RSS_L3_DST_ONLY, "l3-dst-only") \
+ _(63, ETH_RSS_L3_SRC_ONLY, "l3-src-only")
+
format_function_t format_dpdk_device_name;
format_function_t format_dpdk_device;
-format_function_t format_dpdk_tx_dma_trace;
-format_function_t format_dpdk_rx_dma_trace;
+format_function_t format_dpdk_device_errors;
+format_function_t format_dpdk_tx_trace;
+format_function_t format_dpdk_rx_trace;
format_function_t format_dpdk_rte_mbuf;
format_function_t format_dpdk_rx_rte_mbuf;
-unformat_function_t unformat_socket_mem;
+format_function_t format_dpdk_flow;
+format_function_t format_dpdk_rss_hf_name;
+format_function_t format_dpdk_rx_offload_caps;
+format_function_t format_dpdk_tx_offload_caps;
+vnet_flow_dev_ops_function_t dpdk_flow_ops_fn;
+
clib_error_t *unformat_rss_fn (unformat_input_t * input, uword * rss_fn);
-clib_error_t *unformat_hqos (unformat_input_t * input,
- dpdk_device_config_hqos_t * hqos);
-uword
-admin_up_down_process (vlib_main_t * vm,
- vlib_node_runtime_t * rt, vlib_frame_t * f);
+struct rte_pci_device *dpdk_get_pci_device (const struct rte_eth_dev_info
+ *info);
+void dpdk_cli_reference (void);
+
+#if CLI_DEBUG
+int dpdk_buffer_validate_trajectory_all (u32 * uninitialized);
+void dpdk_buffer_poison_trajectory_all (void);
+#endif
#endif /* __included_dpdk_h__ */