#define vlib_validate_buffer_enqueue_x2(vm,node,next_index,to_next,n_left_to_next,bi0,bi1,next0,next1) \
do { \
+ ASSERT (bi0 != 0); \
+ ASSERT (bi1 != 0); \
int enqueue_code = (next0 != next_index) + 2*(next1 != next_index); \
\
if (PREDICT_FALSE (enqueue_code != 0)) \
#define vlib_validate_buffer_enqueue_x4(vm,node,next_index,to_next,n_left_to_next,bi0,bi1,bi2,bi3,next0,next1,next2,next3) \
do { \
+ ASSERT (bi0 != 0); \
+ ASSERT (bi1 != 0); \
+ ASSERT (bi2 != 0); \
+ ASSERT (bi3 != 0); \
/* After the fact: check the [speculative] enqueue to "next" */ \
u32 fix_speculation = (next_index ^ next0) | (next_index ^ next1) \
| (next_index ^ next2) | (next_index ^ next3); \
*/
#define vlib_validate_buffer_enqueue_x1(vm,node,next_index,to_next,n_left_to_next,bi0,next0) \
do { \
+ ASSERT (bi0 != 0); \
if (PREDICT_FALSE (next0 != next_index)) \
{ \
vlib_put_next_frame (vm, node, next_index, n_left_to_next + 1); \
vlib_buffer_enqueue_to_next (vlib_main_t * vm, vlib_node_runtime_t * node,
u32 * buffers, u16 * nexts, uword count)
{
- u32 *to_next, n_left_to_next, max;
- u16 next_index;
-
- next_index = nexts[0];
- vlib_get_next_frame (vm, node, next_index, to_next, n_left_to_next);
- max = clib_min (n_left_to_next, count);
-
- while (count)
- {
- u32 n_enqueued;
- if ((nexts[0] != next_index) || n_left_to_next == 0)
- {
- vlib_put_next_frame (vm, node, next_index, n_left_to_next);
- next_index = nexts[0];
- vlib_get_next_frame (vm, node, next_index, to_next, n_left_to_next);
- max = clib_min (n_left_to_next, count);
- }
-#if defined(CLIB_HAVE_VEC512)
- u16x32 next32 = u16x32_load_unaligned (nexts);
- next32 = (next32 == u16x32_splat (next32[0]));
- u64 bitmap = u16x32_msb_mask (next32);
- n_enqueued = count_trailing_zeros (~bitmap);
-#elif defined(CLIB_HAVE_VEC256)
- u16x16 next16 = u16x16_load_unaligned (nexts);
- next16 = (next16 == u16x16_splat (next16[0]));
- u64 bitmap = u8x32_msb_mask ((u8x32) next16);
- n_enqueued = count_trailing_zeros (~bitmap) / 2;
-#elif defined(CLIB_HAVE_VEC128)
- u16x8 next8 = u16x8_load_unaligned (nexts);
- next8 = (next8 == u16x8_splat (next8[0]));
- u64 bitmap = u8x16_msb_mask ((u8x16) next8);
- n_enqueued = count_trailing_zeros (~bitmap) / 2;
-#else
- u16 x = 0;
- x |= next_index ^ nexts[1];
- x |= next_index ^ nexts[2];
- x |= next_index ^ nexts[3];
- n_enqueued = (x == 0) ? 4 : 1;
-#endif
-
- if (PREDICT_FALSE (n_enqueued > max))
- n_enqueued = max;
-
-#ifdef CLIB_HAVE_VEC512
- if (n_enqueued >= 32)
- {
- clib_memcpy (to_next, buffers, 32 * sizeof (u32));
- nexts += 32;
- to_next += 32;
- buffers += 32;
- n_left_to_next -= 32;
- count -= 32;
- max -= 32;
- continue;
- }
-#endif
-
-#ifdef CLIB_HAVE_VEC256
- if (n_enqueued >= 16)
- {
- clib_memcpy (to_next, buffers, 16 * sizeof (u32));
- nexts += 16;
- to_next += 16;
- buffers += 16;
- n_left_to_next -= 16;
- count -= 16;
- max -= 16;
- continue;
- }
-#endif
-
-#ifdef CLIB_HAVE_VEC128
- if (n_enqueued >= 8)
- {
- clib_memcpy (to_next, buffers, 8 * sizeof (u32));
- nexts += 8;
- to_next += 8;
- buffers += 8;
- n_left_to_next -= 8;
- count -= 8;
- max -= 8;
- continue;
- }
-#endif
-
- if (n_enqueued >= 4)
- {
- clib_memcpy (to_next, buffers, 4 * sizeof (u32));
- nexts += 4;
- to_next += 4;
- buffers += 4;
- n_left_to_next -= 4;
- count -= 4;
- max -= 4;
- continue;
- }
+ vlib_buffer_enqueue_to_next_fn_t *fn;
+ fn = vlib_buffer_func_main.buffer_enqueue_to_next_fn;
+ (fn) (vm, node, buffers, nexts, count);
+}
- /* copy */
- to_next[0] = buffers[0];
+static_always_inline void
+vlib_buffer_enqueue_to_single_next (vlib_main_t * vm,
+ vlib_node_runtime_t * node, u32 * buffers,
+ u16 next_index, u32 count)
+{
+ vlib_buffer_enqueue_to_single_next_fn_t *fn;
+ fn = vlib_buffer_func_main.buffer_enqueue_to_single_next_fn;
+ (fn) (vm, node, buffers, next_index, count);
+}
- /* next */
- nexts += 1;
- to_next += 1;
- buffers += 1;
- n_left_to_next -= 1;
- count -= 1;
- max -= 1;
- }
- vlib_put_next_frame (vm, node, next_index, n_left_to_next);
+static_always_inline u32
+vlib_buffer_enqueue_to_thread (vlib_main_t * vm, u32 frame_queue_index,
+ u32 * buffer_indices, u16 * thread_indices,
+ u32 n_packets, int drop_on_congestion)
+{
+ vlib_buffer_enqueue_to_thread_fn_t *fn;
+ fn = vlib_buffer_func_main.buffer_enqueue_to_thread_fn;
+ return (fn) (vm, frame_queue_index, buffer_indices, thread_indices,
+ n_packets, drop_on_congestion);
}
#endif /* included_vlib_buffer_node_h */