X-Git-Url: https://gerrit.fd.io/r/gitweb?a=blobdiff_plain;f=src%2Fvlib%2Fbuffer.h;h=02b170907ba0ef727d39fc95702505266afb676a;hb=d3d424180fcd98561d656d5c2189e9e59ef2b2b9;hp=5f1e62f08c97deaeedfe9e434cb24389ec7f09ad;hpb=7cd468a3d7dee7d6c92f69a0bb7061ae208ec727;p=vpp.git diff --git a/src/vlib/buffer.h b/src/vlib/buffer.h index 5f1e62f08c9..02b170907ba 100644 --- a/src/vlib/buffer.h +++ b/src/vlib/buffer.h @@ -44,23 +44,22 @@ #include #include #include +#include #include /* for vlib_error_t */ -#if DPDK > 0 -#include -#define VLIB_BUFFER_DATA_SIZE (2048) -#define VLIB_BUFFER_PRE_DATA_SIZE RTE_PKTMBUF_HEADROOM -#else #include /* for __PRE_DATA_SIZE */ -#define VLIB_BUFFER_DATA_SIZE (512) +#define VLIB_BUFFER_DATA_SIZE (2048) #define VLIB_BUFFER_PRE_DATA_SIZE __PRE_DATA_SIZE -#endif -#if defined (CLIB_HAVE_VEC128) || defined (__aarch64__) -typedef u8x16 vlib_copy_unit_t; -#else -typedef u64 vlib_copy_unit_t; -#endif +/* Minimum buffer chain segment size. Does not apply to last buffer in chain. + Dataplane code can safely asume that specified amount of data is not split + into 2 chained buffers */ +#define VLIB_BUFFER_MIN_CHAIN_SEG_SIZE (128) + +/* Amount of head buffer data copied to each replica head buffer */ +#define VLIB_BUFFER_CLONE_HEAD_SIZE (256) + +typedef u8 vlib_buffer_free_list_index_t; /** \file vlib buffer structure definition and a few select @@ -69,10 +68,42 @@ typedef u64 vlib_copy_unit_t; of typing to make it so. */ +/** + * Buffer Flags + */ +#define foreach_vlib_buffer_flag \ + _( 0, NON_DEFAULT_FREELIST, "non-default-fl") \ + _( 1, IS_TRACED, 0) \ + _( 2, NEXT_PRESENT, 0) \ + _( 3, TOTAL_LENGTH_VALID, 0) \ + _( 4, EXT_HDR_VALID, "ext-hdr-valid") + +/* NOTE: only buffer generic flags should be defined here, please consider + using user flags. i.e. src/vnet/buffer.h */ + +enum +{ +#define _(bit, name, v) VLIB_BUFFER_##name = (1 << (bit)), + foreach_vlib_buffer_flag +#undef _ +}; + +enum +{ +#define _(bit, name, v) VLIB_BUFFER_LOG2_##name = (bit), + foreach_vlib_buffer_flag +#undef _ +}; + + /* User defined buffer flags. */ +#define LOG2_VLIB_BUFFER_FLAG_USER(n) (32 - (n)) +#define VLIB_BUFFER_FLAG_USER(n) (1 << LOG2_VLIB_BUFFER_FLAG_USER(n)) + /* VLIB buffer representation. */ typedef struct { CLIB_CACHE_LINE_ALIGN_MARK (cacheline0); + STRUCT_MARK (template_start); /* Offset within data[] that we are currently processing. If negative current header points into predata area. */ i16 current_data; /**< signed offset in data[], pre_data[] @@ -83,58 +114,35 @@ typedef struct the end of this buffer. */ u32 flags; /**< buffer flags: +
VLIB_BUFFER_FREE_LIST_INDEX_MASK: bits used to store free list index,
VLIB_BUFFER_IS_TRACED: trace this buffer.
VLIB_BUFFER_NEXT_PRESENT: this is a multi-chunk buffer.
VLIB_BUFFER_TOTAL_LENGTH_VALID: as it says -
VLIB_BUFFER_REPL_FAIL: packet replication failure -
VLIB_BUFFER_RECYCLE: as it says -
VLIB_BUFFER_FLOW_REPORT: buffer is a flow report, +
VLIB_BUFFER_EXT_HDR_VALID: buffer contains valid external buffer manager header, set to avoid adding it to a flow report
VLIB_BUFFER_FLAG_USER(n): user-defined bit N */ -#define VLIB_BUFFER_IS_TRACED (1 << 0) -#define VLIB_BUFFER_LOG2_NEXT_PRESENT (1) -#define VLIB_BUFFER_NEXT_PRESENT (1 << VLIB_BUFFER_LOG2_NEXT_PRESENT) -#define VLIB_BUFFER_IS_RECYCLED (1 << 2) -#define VLIB_BUFFER_TOTAL_LENGTH_VALID (1 << 3) -#define VLIB_BUFFER_REPL_FAIL (1 << 4) -#define VLIB_BUFFER_RECYCLE (1 << 5) -#define VLIB_BUFFER_FLOW_REPORT (1 << 6) - /* User defined buffer flags. */ -#define LOG2_VLIB_BUFFER_FLAG_USER(n) (32 - (n)) -#define VLIB_BUFFER_FLAG_USER(n) (1 << LOG2_VLIB_BUFFER_FLAG_USER(n)) + u32 flow_id; /**< Generic flow identifier */ - u32 free_list_index; /**< Buffer free list that this buffer was - allocated from and will be freed to. - */ - - u32 total_length_not_including_first_buffer; - /**< Only valid for first buffer in chain. Current length plus - total length given here give total number of bytes in buffer chain. - */ u32 next_buffer; /**< Next buffer for this linked-list of buffers. Only valid if VLIB_BUFFER_NEXT_PRESENT flag is set. */ - vlib_error_t error; /**< Error code for buffers to be enqueued - to error handler. - */ + STRUCT_MARK (template_end); + u32 current_config_index; /**< Used by feature subgraph arcs to visit enabled feature nodes */ - - u8 feature_arc_index; /**< Used to identify feature arcs by intermediate - feature node + vlib_error_t error; /**< Error code for buffers to be enqueued + to error handler. */ + u8 n_add_refs; /**< Number of additional references to this buffer. */ - u8 dont_waste_me[3]; /**< Available space in the (precious) - first 32 octets of buffer metadata - Before allocating any of it, discussion required! - */ + u8 buffer_pool_index; /**< index of buffer pool this buffer belongs. */ - u32 opaque[8]; /**< Opaque data used by sub-graphs for their own purposes. + u32 opaque[10]; /**< Opaque data used by sub-graphs for their own purposes. See .../vnet/vnet/buffer.h */ CLIB_CACHE_LINE_ALIGN_MARK (cacheline1); @@ -143,7 +151,16 @@ typedef struct if VLIB_PACKET_IS_TRACED flag is set. */ u32 recycle_count; /**< Used by L2 path recycle code */ - u32 opaque2[14]; /**< More opaque data, currently unused */ + + u32 total_length_not_including_first_buffer; + /**< Only valid for first buffer in chain. Current length plus + total length given here give total number of bytes in buffer chain. + */ + vlib_buffer_free_list_index_t free_list_index; /** < only used if + VLIB_BUFFER_NON_DEFAULT_FREELIST + flag is set */ + u8 align_pad[3]; /**< available */ + u32 opaque2[12]; /**< More opaque data, see ../vnet/vnet/buffer.h */ /***** end of second cache line */ CLIB_CACHE_LINE_ALIGN_MARK (cacheline2); @@ -168,12 +185,8 @@ typedef struct */ #define vlib_prefetch_buffer_header(b,type) CLIB_PREFETCH (b, 64, type) - -always_inline vlib_buffer_t * -vlib_buffer_next_contiguous (vlib_buffer_t * b, u32 buffer_bytes) -{ - return (void *) (b + 1) + buffer_bytes; -} +#define vlib_prefetch_buffer_data(b,type) \ + CLIB_PREFETCH (vlib_buffer_get_current(b), CLIB_CACHE_LINE_BYTES, type) always_inline void vlib_buffer_struct_is_sane (vlib_buffer_t * b) @@ -184,6 +197,12 @@ vlib_buffer_struct_is_sane (vlib_buffer_t * b) ASSERT (b->pre_data + VLIB_BUFFER_PRE_DATA_SIZE == b->data); } +always_inline uword +vlib_buffer_get_va (vlib_buffer_t * b) +{ + return pointer_to_uword (b->data); +} + /** \brief Get pointer to current data to process @param b - (vlib_buffer_t *) pointer to the buffer @@ -198,6 +217,12 @@ vlib_buffer_get_current (vlib_buffer_t * b) return b->data + b->current_data; } +always_inline uword +vlib_buffer_get_current_va (vlib_buffer_t * b) +{ + return vlib_buffer_get_va (b) + b->current_data; +} + /** \brief Advance current data pointer by the supplied (signed!) amount @param b - (vlib_buffer_t *) pointer to the buffer @@ -209,6 +234,21 @@ vlib_buffer_advance (vlib_buffer_t * b, word l) ASSERT (b->current_length >= l); b->current_data += l; b->current_length -= l; + + ASSERT ((b->flags & VLIB_BUFFER_NEXT_PRESENT) == 0 || + b->current_length >= VLIB_BUFFER_MIN_CHAIN_SEG_SIZE); +} + +/** \brief Check if there is enough space in buffer to advance + + @param b - (vlib_buffer_t *) pointer to the buffer + @param l - (word) size to check + @return - 0 if there is less space than 'l' in buffer +*/ +always_inline u8 +vlib_buffer_has_space (vlib_buffer_t * b, word l) +{ + return b->current_length >= l; } /** \brief Reset current header & length to state they were in when @@ -246,6 +286,74 @@ vlib_get_buffer_opaque2 (vlib_buffer_t * b) return (void *) b->opaque2; } +/** \brief Get pointer to the end of buffer's data + * @param b pointer to the buffer + * @return pointer to tail of packet's data + */ +always_inline u8 * +vlib_buffer_get_tail (vlib_buffer_t * b) +{ + return b->data + b->current_data + b->current_length; +} + +/** \brief Append uninitialized data to buffer + * @param b pointer to the buffer + * @param size number of uninitialized bytes + * @return pointer to beginning of uninitialized data + */ +always_inline void * +vlib_buffer_put_uninit (vlib_buffer_t * b, u16 size) +{ + void *p = vlib_buffer_get_tail (b); + /* XXX make sure there's enough space */ + b->current_length += size; + return p; +} + +/** \brief Prepend uninitialized data to buffer + * @param b pointer to the buffer + * @param size number of uninitialized bytes + * @return pointer to beginning of uninitialized data + */ +always_inline void * +vlib_buffer_push_uninit (vlib_buffer_t * b, u8 size) +{ + ASSERT (b->current_data + VLIB_BUFFER_PRE_DATA_SIZE >= size); + b->current_data -= size; + b->current_length += size; + + return vlib_buffer_get_current (b); +} + +/** \brief Make head room, typically for packet headers + * @param b pointer to the buffer + * @param size number of head room bytes + * @return pointer to start of buffer (current data) + */ +always_inline void * +vlib_buffer_make_headroom (vlib_buffer_t * b, u8 size) +{ + ASSERT (b->current_data + VLIB_BUFFER_PRE_DATA_SIZE >= size); + b->current_data += size; + return vlib_buffer_get_current (b); +} + +/** \brief Retrieve bytes from buffer head + * @param b pointer to the buffer + * @param size number of bytes to pull + * @return pointer to start of buffer (current data) + */ +always_inline void * +vlib_buffer_pull (vlib_buffer_t * b, u8 size) +{ + if (b->current_length + VLIB_BUFFER_PRE_DATA_SIZE < size) + return 0; + + void *data = vlib_buffer_get_current (b); + vlib_buffer_advance (b, size); + return data; +} + /* Forward declaration. */ struct vlib_main_t; @@ -256,28 +364,22 @@ typedef struct vlib_buffer_free_list_t vlib_buffer_t buffer_init_template; /* Our index into vlib_main_t's buffer_free_list_pool. */ - u32 index; + vlib_buffer_free_list_index_t index; /* Number of data bytes for buffers in this free list. */ u32 n_data_bytes; - /* Number of buffers to allocate when we need to allocate new buffers - from physmem heap. */ - u32 min_n_buffers_each_physmem_alloc; + /* Number of buffers to allocate when we need to allocate new buffers */ + u32 min_n_buffers_each_alloc; /* Total number of buffers allocated from this free list. */ u32 n_alloc; - /* Vector of free buffers. Each element is a byte offset into I/O heap. - Aligned vectors always has naturally aligned vlib_copy_unit_t sized chunks - of buffer indices. Unaligned vector has any left over. This is meant to - speed up copy routines. */ - u32 *aligned_buffers, *unaligned_buffers; + /* Vector of free buffers. Each element is a byte offset into I/O heap. */ + u32 *buffers; - /* Memory chunks allocated for this free list - recorded here so they can be freed when free list - is deleted. */ - void **buffer_memory_allocated; + /* index of buffer pool used to get / put buffers */ + u8 buffer_pool_index; /* Free list name. */ u8 *name; @@ -288,25 +390,60 @@ typedef struct vlib_buffer_free_list_t struct vlib_buffer_free_list_t * fl, u32 * buffers, u32 n_buffers); - /* Callback function to announce that buffers have been - added to the freelist */ - void (*buffers_added_to_freelist_function) - (struct vlib_main_t * vm, struct vlib_buffer_free_list_t * fl); - uword buffer_init_function_opaque; } __attribute__ ((aligned (16))) vlib_buffer_free_list_t; +typedef uword (vlib_buffer_fill_free_list_cb_t) (struct vlib_main_t * vm, + vlib_buffer_free_list_t * fl, + uword min_free_buffers); +typedef void (vlib_buffer_free_cb_t) (struct vlib_main_t * vm, u32 * buffers, + u32 n_buffers); +typedef void (vlib_buffer_free_no_next_cb_t) (struct vlib_main_t * vm, + u32 * buffers, u32 n_buffers); + typedef struct { + vlib_buffer_fill_free_list_cb_t *vlib_buffer_fill_free_list_cb; + vlib_buffer_free_cb_t *vlib_buffer_free_cb; + vlib_buffer_free_no_next_cb_t *vlib_buffer_free_no_next_cb; + void (*vlib_packet_template_init_cb) (struct vlib_main_t * vm, void *t, + void *packet_data, + uword n_packet_data_bytes, + uword + min_n_buffers_each_physmem_alloc, + u8 * name); + void (*vlib_buffer_delete_free_list_cb) (struct vlib_main_t * vm, + vlib_buffer_free_list_index_t + free_list_index); +} vlib_buffer_callbacks_t; + +extern vlib_buffer_callbacks_t *vlib_buffer_callbacks; + +typedef struct +{ + CLIB_CACHE_LINE_ALIGN_MARK (cacheline0); + uword start; + uword size; + uword log2_page_size; + u32 physmem_map_index; + u32 buffer_size; + u32 *buffers; + clib_spinlock_t lock; +} vlib_buffer_pool_t; + +typedef struct +{ + CLIB_CACHE_LINE_ALIGN_MARK (cacheline0); + /* Virtual memory address and size of buffer memory, used for calculating + buffer index */ + uword buffer_mem_start; + uword buffer_mem_size; + vlib_buffer_pool_t *buffer_pools; + /* Buffer free callback, for subversive activities */ - u32 (*buffer_free_callback) (struct vlib_main_t * vm, - u32 * buffers, - u32 n_buffers, u32 follow_buffer_next); - /* Pool of buffer free lists. - Multiple free lists exist for packet generator which uses - separate free lists for each packet stream --- so as to avoid - initializing static data for each packet generated. */ - vlib_buffer_free_list_t *buffer_free_list_pool; + u32 (*buffer_free_callback) (struct vlib_main_t * vm, + u32 * buffers, + u32 n_buffers, u32 follow_buffer_next); #define VLIB_BUFFER_DEFAULT_FREE_LIST_INDEX (0) #define VLIB_BUFFER_DEFAULT_FREE_LIST_BYTES VLIB_BUFFER_DATA_SIZE @@ -319,16 +456,27 @@ typedef struct If buffer index is not in hash table then this buffer has never been allocated. */ uword *buffer_known_hash; + clib_spinlock_t buffer_known_hash_lockp; - /* List of free-lists needing Blue Light Special announcements */ - vlib_buffer_free_list_t **announce_list; - - /* Vector of rte_mempools per socket */ -#if DPDK == 1 - struct rte_mempool **pktmbuf_pools; -#endif + /* Callbacks */ + vlib_buffer_callbacks_t cb; + int callbacks_registered; } vlib_buffer_main_t; +extern vlib_buffer_main_t buffer_main; + +static_always_inline vlib_buffer_pool_t * +vlib_buffer_pool_get (u8 buffer_pool_index) +{ + vlib_buffer_main_t *bm = &buffer_main; + return vec_elt_at_index (bm->buffer_pools, buffer_pool_index); +} + +u8 vlib_buffer_register_physmem_map (struct vlib_main_t * vm, + u32 physmem_map_index); + +clib_error_t *vlib_buffer_main_init (struct vlib_main_t *vm); + typedef struct { struct vlib_main_t *vlib_main; @@ -350,7 +498,7 @@ typedef struct u32 next_index; /* Free list to use to allocate new buffers. */ - u32 free_list_index; + vlib_buffer_free_list_index_t free_list_index; } tx; struct @@ -385,11 +533,6 @@ serialize_vlib_buffer_n_bytes (serialize_main_t * m) vec_len (s->overflow_buffer); } -#if DPDK > 0 -#define rte_mbuf_from_vlib_buffer(x) (((struct rte_mbuf *)x) - 1) -#define vlib_buffer_from_rte_mbuf(x) ((vlib_buffer_t *)(x+1)) -#endif - /* */ @@ -401,13 +544,33 @@ serialize_vlib_buffer_n_bytes (serialize_main_t * m) #define VLIB_BUFFER_TRACE_TRAJECTORY 0 #if VLIB_BUFFER_TRACE_TRAJECTORY > 0 -#define VLIB_BUFFER_TRACE_TRAJECTORY_INIT(b) (b)->pre_data[0]=0 +extern void (*vlib_buffer_trace_trajectory_cb) (vlib_buffer_t * b, u32 index); +extern void (*vlib_buffer_trace_trajectory_init_cb) (vlib_buffer_t * b); +extern void vlib_buffer_trace_trajectory_init (vlib_buffer_t * b); +#define VLIB_BUFFER_TRACE_TRAJECTORY_INIT(b) \ + vlib_buffer_trace_trajectory_init (b); #else #define VLIB_BUFFER_TRACE_TRAJECTORY_INIT(b) #endif /* VLIB_BUFFER_TRACE_TRAJECTORY */ #endif /* included_vlib_buffer_h */ +#define VLIB_BUFFER_REGISTER_CALLBACKS(x,...) \ + __VA_ARGS__ vlib_buffer_callbacks_t __##x##_buffer_callbacks; \ +static void __vlib_add_buffer_callbacks_t_##x (void) \ + __attribute__((__constructor__)) ; \ +static void __vlib_add_buffer_callbacks_t_##x (void) \ +{ \ + if (vlib_buffer_callbacks) \ + clib_panic ("vlib buffer callbacks already registered"); \ + vlib_buffer_callbacks = &__##x##_buffer_callbacks; \ +} \ +static void __vlib_rm_buffer_callbacks_t_##x (void) \ + __attribute__((__destructor__)) ; \ +static void __vlib_rm_buffer_callbacks_t_##x (void) \ +{ vlib_buffer_callbacks = 0; } \ +__VA_ARGS__ vlib_buffer_callbacks_t __##x##_buffer_callbacks + /* * fd.io coding-style-patch-verification: ON *