#include <rte_dev.h>
#include <rte_log.h>
#include <rte_memory.h>
-#include <rte_memcpy.h>
#include <rte_memzone.h>
#include <rte_tailq.h>
#include <rte_eal.h>
#include <rte_ring.h>
#include <rte_mempool.h>
#include <rte_mbuf.h>
-#include <rte_kni.h>
#include <rte_virtio_net.h>
-#include <rte_pci_dev_ids.h>
#include <rte_version.h>
+#include <rte_eth_bond.h>
+#include <rte_sched.h>
#include <vnet/unix/pcap.h>
-#include <vnet/devices/virtio/vhost-user.h>
#if CLIB_DEBUG > 0
#define always_inline static inline
#define always_inline static inline __attribute__ ((__always_inline__))
#endif
-#define MBUF_SIZE (2048 + sizeof(struct rte_mbuf) + RTE_PKTMBUF_HEADROOM)
-#define NB_MBUF (32<<10)
+#include <vlib/pci/pci.h>
-vnet_device_class_t dpdk_device_class;
-vlib_node_registration_t dpdk_input_node;
-vlib_node_registration_t dpdk_io_input_node;
-vlib_node_registration_t handoff_dispatch_node;
+#define NB_MBUF (16<<10)
-typedef enum {
- VNET_DPDK_DEV_ETH = 1, /* Standard DPDK PMD driver */
- VNET_DPDK_DEV_KNI, /* Kernel NIC Interface */
- VNET_DPDK_DEV_VHOST_USER,
- VNET_DPDK_DEV_UNKNOWN, /* must be last */
-} dpdk_device_type_t;
+extern vnet_device_class_t dpdk_device_class;
+extern vlib_node_registration_t dpdk_input_node;
+extern vlib_node_registration_t handoff_dispatch_node;
#define foreach_dpdk_pmd \
_ ("rte_nicvf_pmd", THUNDERX) \
_ ("rte_i40e_pmd", I40E) \
_ ("rte_i40evf_pmd", I40EVF) \
_ ("rte_virtio_pmd", VIRTIO) \
- _ ("rte_vice_pmd", VICE) \
_ ("rte_enic_pmd", ENIC) \
_ ("rte_vmxnet3_pmd", VMXNET3) \
_ ("AF_PACKET PMD", AF_PACKET) \
+ _ ("rte_bond_pmd", BOND) \
_ ("rte_pmd_fm10k", FM10K) \
- _ ("rte_cxgbe_pmd", CXGBE)
+ _ ("rte_cxgbe_pmd", CXGBE) \
+ _ ("rte_dpaa2_dpni", DPAA2)
-typedef enum {
+typedef enum
+{
VNET_DPDK_PMD_NONE,
#define _(s,f) VNET_DPDK_PMD_##f,
foreach_dpdk_pmd
#undef _
-#ifdef NETMAP
- VNET_DPDK_PMD_NETMAP,
-#endif
- VNET_DPDK_PMD_UNKNOWN, /* must be last */
+ VNET_DPDK_PMD_UNKNOWN, /* must be last */
} dpdk_pmd_t;
-typedef enum {
+typedef enum
+{
VNET_DPDK_PORT_TYPE_ETH_1G,
VNET_DPDK_PORT_TYPE_ETH_10G,
VNET_DPDK_PORT_TYPE_ETH_40G,
+ VNET_DPDK_PORT_TYPE_ETH_BOND,
VNET_DPDK_PORT_TYPE_ETH_SWITCH,
-#ifdef NETMAP
- VNET_DPDK_PORT_TYPE_NETMAP,
-#endif
VNET_DPDK_PORT_TYPE_AF_PACKET,
VNET_DPDK_PORT_TYPE_UNKNOWN,
} dpdk_port_type_t;
-typedef struct {
+typedef struct
+{
f64 deadline;
- vlib_frame_t * frame;
+ vlib_frame_t *frame;
} dpdk_frame_t;
#define DPDK_EFD_MAX_DISCARD_RATE 10
-typedef struct {
+typedef struct
+{
u16 last_burst_sz;
u16 max_burst_sz;
u32 full_frames_cnt;
u32 total_packet_cnt;
} dpdk_efd_agent_t;
-typedef struct {
- int callfd;
- int kickfd;
- int errfd;
-#if RTE_VERSION >= RTE_VERSION_NUM(2, 2, 0, 0)
- int enabled;
-#endif
- u32 callfd_idx;
- u32 n_since_last_int;
- f64 int_deadline;
-} dpdk_vu_vring;
-
-typedef struct {
- u32 is_up;
- u32 unix_fd;
- u32 unix_file_index;
- u32 client_fd;
- char sock_filename[256];
- int sock_errno;
- u8 sock_is_server;
- u8 active;
-
- u64 feature_mask;
- u32 num_vrings;
-#if RTE_VERSION >= RTE_VERSION_NUM(2, 2, 0, 0)
- dpdk_vu_vring vrings[VHOST_MAX_QUEUE_PAIRS * 2];
-#else
- dpdk_vu_vring vrings[2];
-#endif
- u64 region_addr[VHOST_MEMORY_MAX_NREGIONS];
- u32 region_fd[VHOST_MEMORY_MAX_NREGIONS];
-} dpdk_vu_intf_t;
-
-typedef void (*dpdk_flowcontrol_callback_t) (vlib_main_t *vm,
- u32 hw_if_index,
- u32 n_packets);
+typedef void (*dpdk_flowcontrol_callback_t) (vlib_main_t * vm,
+ u32 hw_if_index, u32 n_packets);
/*
* The header for the tx_vector in dpdk_device_t.
* Head and tail are indexes into the tx_vector and are of type
* u64 so they never overflow.
*/
-typedef struct {
+typedef struct
+{
u64 tx_head;
u64 tx_tail;
} tx_ring_hdr_t;
-typedef struct {
- CLIB_CACHE_LINE_ALIGN_MARK(cacheline0);
- volatile u32 *lockp;
+typedef struct
+{
+ struct rte_ring *swq;
+
+ u64 hqos_field0_slabmask;
+ u32 hqos_field0_slabpos;
+ u32 hqos_field0_slabshr;
+ u64 hqos_field1_slabmask;
+ u32 hqos_field1_slabpos;
+ u32 hqos_field1_slabshr;
+ u64 hqos_field2_slabmask;
+ u32 hqos_field2_slabpos;
+ u32 hqos_field2_slabshr;
+ u32 hqos_tc_table[64];
+} dpdk_device_hqos_per_worker_thread_t;
+
+typedef struct
+{
+ struct rte_ring **swq;
+ struct rte_mbuf **pkts_enq;
+ struct rte_mbuf **pkts_deq;
+ struct rte_sched_port *hqos;
+ u32 hqos_burst_enq;
+ u32 hqos_burst_deq;
+ u32 pkts_enq_len;
+ u32 swq_pos;
+} dpdk_device_hqos_per_hqos_thread_t;
+
+typedef struct
+{
+ CLIB_CACHE_LINE_ALIGN_MARK (cacheline0);
+ volatile u32 **lockp;
/* Instance ID */
u32 device_index;
u32 per_interface_next_index;
/* dpdk rte_mbuf rx and tx vectors, VLIB_FRAME_SIZE */
- struct rte_mbuf *** tx_vectors; /* one per worker thread */
- struct rte_mbuf *** rx_vectors;
+ struct rte_mbuf ***tx_vectors; /* one per worker thread */
+ struct rte_mbuf ***rx_vectors;
/* vector of traced contexts, per device */
- u32 * d_trace_buffers;
-
- /* per-worker destination frame queue */
- dpdk_frame_t * frames;
+ u32 *d_trace_buffers;
- dpdk_device_type_t dev_type:8;
dpdk_pmd_t pmd:8;
i8 cpu_socket;
- u8 admin_up;
- u8 promisc;
+ u16 flags;
+#define DPDK_DEVICE_FLAG_ADMIN_UP (1 << 0)
+#define DPDK_DEVICE_FLAG_PROMISC (1 << 1)
+#define DPDK_DEVICE_FLAG_PMD (1 << 2)
- CLIB_CACHE_LINE_ALIGN_MARK(cacheline1);
+#define DPDK_DEVICE_FLAG_HAVE_SUBIF (1 << 5)
+#define DPDK_DEVICE_FLAG_HQOS (1 << 6)
+
+ u16 nb_tx_desc;
+ CLIB_CACHE_LINE_ALIGN_MARK (cacheline1);
+
+ u8 *interface_name_suffix;
+
+ /* number of sub-interfaces */
+ u16 num_subifs;
/* PMD related */
u16 tx_q_used;
u16 rx_q_used;
u16 nb_rx_desc;
- u16 nb_tx_desc;
- u16 * cpu_socket_id_by_queue;
+ u16 *cpu_socket_id_by_queue;
struct rte_eth_conf port_conf;
struct rte_eth_txconf tx_conf;
- /* KNI related */
- struct rte_kni *kni;
- u8 kni_port_id;
-
- /* vhost-user related */
- u32 vu_if_id;
- struct virtio_net vu_vhost_dev;
- u32 vu_is_running;
- dpdk_vu_intf_t *vu_intf;
+ /* HQoS related */
+ dpdk_device_hqos_per_worker_thread_t *hqos_wt;
+ dpdk_device_hqos_per_hqos_thread_t *hqos_ht;
/* af_packet */
u8 af_packet_port_id;
struct rte_eth_stats stats;
struct rte_eth_stats last_stats;
- struct rte_eth_xstats * xstats;
+ struct rte_eth_stats last_cleared_stats;
+ struct rte_eth_xstat *xstats;
+ struct rte_eth_xstat *last_cleared_xstats;
f64 time_last_stats_update;
dpdk_port_type_t port_type;
dpdk_efd_agent_t efd_agent;
} dpdk_device_t;
-#define MAX_NELTS 32
-typedef struct {
- CLIB_CACHE_LINE_ALIGN_MARK(cacheline0);
- u64 head;
- u64 head_hint;
- u64 tail;
- u32 n_in_use;
- u32 nelts;
- u32 written;
- u32 threshold;
- i32 n_vectors[MAX_NELTS];
-} frame_queue_trace_t;
-
-#define DPDK_TX_RING_SIZE (4 * 1024)
+#define DPDK_STATS_POLL_INTERVAL (10.0)
+#define DPDK_MIN_STATS_POLL_INTERVAL (0.001) /* 1msec */
-#define DPDK_STATS_POLL_INTERVAL 10.0
-#define DPDK_LINK_POLL_INTERVAL 3.0
+#define DPDK_LINK_POLL_INTERVAL (3.0)
+#define DPDK_MIN_LINK_POLL_INTERVAL (0.001) /* 1msec */
-typedef struct {
- CLIB_CACHE_LINE_ALIGN_MARK(cacheline0);
+typedef struct
+{
+ CLIB_CACHE_LINE_ALIGN_MARK (cacheline0);
/* total input packet counter */
u64 aggregate_rx_packets;
} dpdk_worker_t;
-typedef struct {
+typedef struct
+{
+ CLIB_CACHE_LINE_ALIGN_MARK (cacheline0);
+
+ /* total input packet counter */
+ u64 aggregate_rx_packets;
+} dpdk_hqos_thread_t;
+
+typedef struct
+{
u32 device;
u16 queue_id;
} dpdk_device_and_queue_t;
#define DPDK_EFD_DEFAULT_DEVICE_QUEUE_HI_THRESH_PCT 90
#define DPDK_EFD_DEFAULT_CONSEC_FULL_FRAMES_HI_THRESH 6
-typedef struct dpdk_efd_t {
+typedef struct dpdk_efd_t
+{
u16 enabled;
u16 queue_hi_thresh;
u16 consec_full_frames_hi_thresh;
u16 pad;
} dpdk_efd_t;
-typedef struct {
+#ifndef DPDK_HQOS_DBG_BYPASS
+#define DPDK_HQOS_DBG_BYPASS 0
+#endif
- /* Devices */
- dpdk_device_t * devices;
- dpdk_device_and_queue_t ** devices_by_cpu;
+typedef struct dpdk_device_config_hqos_t
+{
+ u32 hqos_thread;
+ u32 hqos_thread_valid;
+
+ u32 swq_size;
+ u32 burst_enq;
+ u32 burst_deq;
+
+ u32 pktfield0_slabpos;
+ u32 pktfield1_slabpos;
+ u32 pktfield2_slabpos;
+ u64 pktfield0_slabmask;
+ u64 pktfield1_slabmask;
+ u64 pktfield2_slabmask;
+ u32 tc_table[64];
+
+ struct rte_sched_port_params port;
+ struct rte_sched_subport_params *subport;
+ struct rte_sched_pipe_params *pipe;
+ uint32_t *pipe_map;
+} dpdk_device_config_hqos_t;
+
+int dpdk_hqos_validate_mask (u64 mask, u32 n);
+void dpdk_device_config_hqos_pipe_profile_default (dpdk_device_config_hqos_t *
+ hqos, u32 pipe_profile_id);
+void dpdk_device_config_hqos_default (dpdk_device_config_hqos_t * hqos);
+clib_error_t *dpdk_port_setup_hqos (dpdk_device_t * xd,
+ dpdk_device_config_hqos_t * hqos);
+void dpdk_hqos_metadata_set (dpdk_device_hqos_per_worker_thread_t * hqos,
+ struct rte_mbuf **pkts, u32 n_pkts);
+
+#define foreach_dpdk_device_config_item \
+ _ (num_rx_queues) \
+ _ (num_tx_queues) \
+ _ (num_rx_desc) \
+ _ (num_tx_desc) \
+ _ (rss_fn)
+
+typedef struct
+{
+ vlib_pci_addr_t pci_addr;
+ u8 is_blacklisted;
+ u8 vlan_strip_offload;
+#define DPDK_DEVICE_VLAN_STRIP_DEFAULT 0
+#define DPDK_DEVICE_VLAN_STRIP_OFF 1
+#define DPDK_DEVICE_VLAN_STRIP_ON 2
+
+#define _(x) uword x;
+ foreach_dpdk_device_config_item
+#undef _
+ clib_bitmap_t * workers;
+ u32 hqos_enabled;
+ dpdk_device_config_hqos_t hqos;
+} dpdk_device_config_t;
- /* per-thread recycle lists */
- u32 ** recycle;
+typedef struct
+{
- /* flow control callback. If 0 then flow control is disabled */
- dpdk_flowcontrol_callback_t flowcontrol_callback;
+ /* Config stuff */
+ u8 **eal_init_args;
+ u8 *eal_init_args_str;
+ u8 *uio_driver_name;
+ u8 no_multi_seg;
+ u8 enable_tcp_udp_checksum;
- /* vlib buffer free list, must be same size as an rte_mbuf */
- u32 vlib_buffer_free_list_index;
+ /* Required config parameters */
+ u8 coremask_set_manually;
+ u8 nchannels_set_manually;
+ u32 coremask;
+ u32 nchannels;
+ u32 num_mbufs;
+ u8 num_kni; /* while kni_init allows u32, port_id in callback fn is only u8 */
/*
* format interface names ala xxxEthernet%d/%d/%d instead of
- * xxxEthernet%x/%x/%x. For VIRL.
+ * xxxEthernet%x/%x/%x.
*/
u8 interface_name_format_decimal;
+ /* per-device config */
+ dpdk_device_config_t default_devconf;
+ dpdk_device_config_t *dev_confs;
+ uword *device_config_index_by_pci_addr;
- /* dpdk worker "threads" */
- dpdk_worker_t * workers;
+} dpdk_config_main_t;
- /* Config stuff */
- u8 ** eal_init_args;
- u8 * eth_if_blacklist;
- u8 * eth_if_whitelist;
- u8 * uio_driver_name;
- u8 no_multi_seg;
+dpdk_config_main_t dpdk_config_main;
- /* Required config parameters */
- u8 coremask_set_manually;
- u8 nchannels_set_manually;
- u32 coremask;
- u32 nchannels;
- u32 num_mbufs;
- u32 use_rss;
- u8 num_kni; /* while kni_init allows u32, port_id in callback fn is only u8 */
+typedef struct
+{
+
+ /* Devices */
+ dpdk_device_t *devices;
+ dpdk_device_and_queue_t **devices_by_cpu;
+ dpdk_device_and_queue_t **devices_by_hqos_cpu;
+
+ /* per-thread recycle lists */
+ u32 **recycle;
+
+ /* buffer flags template, configurable to enable/disable tcp / udp cksum */
+ u32 buffer_flags_template;
+
+ /* flow control callback. If 0 then flow control is disabled */
+ dpdk_flowcontrol_callback_t flowcontrol_callback;
+
+ /* vlib buffer free list, must be same size as an rte_mbuf */
+ u32 vlib_buffer_free_list_index;
+
+ /* dpdk worker "threads" */
+ dpdk_worker_t *workers;
+
+ /* dpdk HQoS "threads" */
+ dpdk_hqos_thread_t *hqos_threads;
/* Ethernet input node index */
u32 ethernet_input_node_index;
- /* dpdk i/o thread initialization barrier */
- volatile u32 io_thread_release;
-
/* pcap tracing [only works if (CLIB_DEBUG > 0)] */
int tx_pcap_enable;
pcap_main_t pcap_main;
- u8 * pcap_filename;
+ u8 *pcap_filename;
u32 pcap_sw_if_index;
u32 pcap_pkts_to_capture;
- /* virtio vhost-user switch */
- u8 use_virtio_vhost;
-
- /* vhost-user coalescence frames config */
- u32 vhost_coalesce_frames;
- f64 vhost_coalesce_time;
-
/* hashes */
- uword * dpdk_device_by_kni_port_id;
- uword * vu_sw_if_index_by_listener_fd;
- uword * vu_sw_if_index_by_sock_fd;
- u32 * vu_inactive_interfaces_device_index;
-
- u32 next_vu_if_id;
+ uword *dpdk_device_by_kni_port_id;
+ uword *vu_sw_if_index_by_listener_fd;
+ uword *vu_sw_if_index_by_sock_fd;
+ u32 *vu_inactive_interfaces_device_index;
/* efd (early-fast-discard) settings */
dpdk_efd_t efd;
*/
u8 admin_up_down_in_progress;
- u8 have_io_threads;
+ u8 use_rss;
/* which cpus are running dpdk-input */
int input_cpu_first_index;
int input_cpu_count;
+ /* which cpus are running I/O TX */
+ int hqos_cpu_first_index;
+ int hqos_cpu_count;
+
+ /* control interval of dpdk link state and stat polling */
+ f64 link_state_poll_interval;
+ f64 stat_poll_interval;
+
+ /* Sleep for this many MS after each device poll */
+ u32 poll_sleep;
+
/* convenience */
- vlib_main_t * vlib_main;
- vnet_main_t * vnet_main;
+ vlib_main_t *vlib_main;
+ vnet_main_t *vnet_main;
+ dpdk_config_main_t *conf;
} dpdk_main_t;
dpdk_main_t dpdk_main;
-typedef enum {
+typedef enum
+{
DPDK_RX_NEXT_IP4_INPUT,
DPDK_RX_NEXT_IP6_INPUT,
DPDK_RX_NEXT_MPLS_INPUT,
DPDK_RX_N_NEXT,
} dpdk_rx_next_t;
+typedef struct
+{
+ u32 buffer_index;
+ u16 device_index;
+ u8 queue_index;
+ struct rte_mbuf mb;
+ /* Copy of VLIB buffer; packet data stored in pre_data. */
+ vlib_buffer_t buffer;
+} dpdk_tx_dma_trace_t;
+
+typedef struct
+{
+ u32 buffer_index;
+ u16 device_index;
+ u16 queue_index;
+ struct rte_mbuf mb;
+ vlib_buffer_t buffer; /* Copy of VLIB buffer; pkt data stored in pre_data. */
+ u8 data[256]; /* First 256 data bytes, used for hexdump */
+} dpdk_rx_dma_trace_t;
+
void vnet_buffer_needs_dpdk_mb (vlib_buffer_t * b);
void dpdk_set_next_node (dpdk_rx_next_t, char *);
-typedef void (*dpdk_io_thread_callback_t) (vlib_main_t *vm);
+clib_error_t *dpdk_set_mac_address (vnet_hw_interface_t * hi, char *address);
+
+clib_error_t *dpdk_set_mc_filter (vnet_hw_interface_t * hi,
+ struct ether_addr mc_addr_vec[], int naddr);
-void dpdk_io_thread (vlib_worker_thread_t * w,
- u32 instances,
- u32 instance_id,
- char *worker_name,
- dpdk_io_thread_callback_t callback);
void dpdk_thread_input (dpdk_main_t * dm, dpdk_device_t * xd);
-clib_error_t * dpdk_port_setup (dpdk_main_t * dm, dpdk_device_t * xd);
+clib_error_t *dpdk_port_setup (dpdk_main_t * dm, dpdk_device_t * xd);
-void dpdk_set_flowcontrol_callback (vlib_main_t *vm,
- dpdk_flowcontrol_callback_t callback);
+void dpdk_set_flowcontrol_callback (vlib_main_t * vm,
+ dpdk_flowcontrol_callback_t callback);
u32 dpdk_interface_tx_vector (vlib_main_t * vm, u32 dev_instance);
-vlib_frame_queue_elt_t * vlib_get_handoff_queue_elt (u32 vlib_worker_index);
-
-u32 dpdk_get_handoff_node_index (void);
+void set_efd_bitmap (u8 * bitmap, u32 value, u32 op);
-void set_efd_bitmap (u8 *bitmap, u32 value, u32 op);
+struct rte_mbuf *dpdk_replicate_packet_mb (vlib_buffer_t * b);
+struct rte_mbuf *dpdk_zerocopy_replicate_packet_mb (vlib_buffer_t * b);
#define foreach_dpdk_error \
_(NONE, "no error") \
_(MPLS_EFD_DROP_PKTS, "MPLS Early Fast Discard rx drops") \
_(VLAN_EFD_DROP_PKTS, "VLAN Early Fast Discard rx drops")
-typedef enum {
+typedef enum
+{
#define _(f,s) DPDK_ERROR_##f,
foreach_dpdk_error
#undef _
- DPDK_N_ERROR,
+ DPDK_N_ERROR,
} dpdk_error_t;
/*
* Increment EFD drop counter
*/
-static_always_inline
-void increment_efd_drop_counter (vlib_main_t * vm, u32 counter_index, u32 count)
+static_always_inline void
+increment_efd_drop_counter (vlib_main_t * vm, u32 counter_index, u32 count)
{
- vlib_node_t *my_n;
+ vlib_node_t *my_n;
- my_n = vlib_get_node (vm, dpdk_input_node.index);
- vm->error_main.counters[my_n->error_heap_index+counter_index] += count;
+ my_n = vlib_get_node (vm, dpdk_input_node.index);
+ vm->error_main.counters[my_n->error_heap_index + counter_index] += count;
}
+int dpdk_set_stat_poll_interval (f64 interval);
+int dpdk_set_link_state_poll_interval (f64 interval);
void dpdk_update_link_state (dpdk_device_t * xd, f64 now);
-void dpdk_efd_update_counters(dpdk_device_t *xd, u32 n_buffers, u16 enabled);
-u32 is_efd_discardable(vlib_thread_main_t *tm,
- vlib_buffer_t * b0,
- struct rte_mbuf *mb);
-
-/* dpdk vhost-user interrupt management */
-u8 dpdk_vhost_user_want_interrupt (dpdk_device_t *xd, int idx);
-void dpdk_vhost_user_send_interrupt (vlib_main_t * vm, dpdk_device_t * xd,
- int idx);
-
-
-static inline u64 vnet_get_aggregate_rx_packets (void)
+void dpdk_device_lock_init (dpdk_device_t * xd);
+void dpdk_device_lock_free (dpdk_device_t * xd);
+void dpdk_efd_update_counters (dpdk_device_t * xd, u32 n_buffers,
+ u16 enabled);
+u32 is_efd_discardable (vlib_thread_main_t * tm, vlib_buffer_t * b0,
+ struct rte_mbuf *mb);
+
+static inline u64
+vnet_get_aggregate_rx_packets (void)
{
- dpdk_main_t * dm = &dpdk_main;
- u64 sum = 0;
- dpdk_worker_t * dw;
+ dpdk_main_t *dm = &dpdk_main;
+ u64 sum = 0;
+ dpdk_worker_t *dw;
- vec_foreach(dw, dm->workers)
- sum += dw->aggregate_rx_packets;
+ vec_foreach (dw, dm->workers) sum += dw->aggregate_rx_packets;
- return sum;
+ return sum;
}
void dpdk_rx_trace (dpdk_main_t * dm,
- vlib_node_runtime_t * node,
- dpdk_device_t * xd,
- u16 queue_id,
- u32 * buffers,
- uword n_buffers);
+ vlib_node_runtime_t * node,
+ dpdk_device_t * xd,
+ u16 queue_id, u32 * buffers, uword n_buffers);
#define EFD_OPERATION_LESS_THAN 0
#define EFD_OPERATION_GREATER_OR_EQUAL 1
-void efd_config(u32 enabled,
- u32 ip_prec, u32 ip_op,
- u32 mpls_exp, u32 mpls_op,
- u32 vlan_cos, u32 vlan_op);
-
-void post_sw_interface_set_flags (vlib_main_t *vm, u32 sw_if_index, u32 flags);
-
-typedef struct vhost_user_memory vhost_user_memory_t;
-
-void dpdk_vhost_user_process_init (void **ctx);
-void dpdk_vhost_user_process_cleanup (void *ctx);
-uword dpdk_vhost_user_process_if (vlib_main_t *vm, dpdk_device_t *xd, void *ctx);
-
-// vhost-user calls
-int dpdk_vhost_user_create_if (vnet_main_t * vnm, vlib_main_t * vm,
- const char * sock_filename,
- u8 is_server,
- u32 * sw_if_index,
- u64 feature_mask,
- u8 renumber, u32 custom_dev_instance);
-int dpdk_vhost_user_modify_if (vnet_main_t * vnm, vlib_main_t * vm,
- const char * sock_filename,
- u8 is_server,
- u32 sw_if_index,
- u64 feature_mask,
- u8 renumber, u32 custom_dev_instance);
-int dpdk_vhost_user_delete_if (vnet_main_t * vnm, vlib_main_t * vm,
- u32 sw_if_index);
-int dpdk_vhost_user_dump_ifs (vnet_main_t * vnm, vlib_main_t * vm,
- vhost_user_intf_details_t **out_vuids);
+void efd_config (u32 enabled,
+ u32 ip_prec, u32 ip_op,
+ u32 mpls_exp, u32 mpls_op, u32 vlan_cos, u32 vlan_op);
+
+void post_sw_interface_set_flags (vlib_main_t * vm, u32 sw_if_index,
+ u32 flags);
u32 dpdk_get_admin_up_down_in_progress (void);
+u32 dpdk_num_mbufs (void);
+
+dpdk_pmd_t dpdk_get_pmd_type (vnet_hw_interface_t * hi);
+
+i8 dpdk_get_cpu_socket (vnet_hw_interface_t * hi);
+
+void *dpdk_input_multiarch_select ();
+void *dpdk_input_rss_multiarch_select ();
+void *dpdk_input_efd_multiarch_select ();
+
+clib_error_t *dpdk_get_hw_interface_stats (u32 hw_if_index,
+ struct rte_eth_stats *dest);
+
+format_function_t format_dpdk_device_name;
+format_function_t format_dpdk_device;
+format_function_t format_dpdk_tx_dma_trace;
+format_function_t format_dpdk_rx_dma_trace;
+format_function_t format_dpdk_rte_mbuf;
+format_function_t format_dpdk_rx_rte_mbuf;
+unformat_function_t unformat_socket_mem;
+clib_error_t *unformat_rss_fn (unformat_input_t * input, uword * rss_fn);
+clib_error_t *unformat_hqos (unformat_input_t * input,
+ dpdk_device_config_hqos_t * hqos);
+
uword
-dpdk_input_rss (vlib_main_t * vm, vlib_node_runtime_t * node, vlib_frame_t * f);
+admin_up_down_process (vlib_main_t * vm,
+ vlib_node_runtime_t * rt, vlib_frame_t * f);
#endif /* __included_dpdk_h__ */
+
+/*
+ * fd.io coding-style-patch-verification: ON
+ *
+ * Local Variables:
+ * eval: (c-set-style "gnu")
+ * End:
+ */