#undef _
};
-always_inline void
-session_tx_trace_buffer (vlib_main_t * vm, vlib_node_runtime_t * node,
- u32 next_index, vlib_buffer_t * b,
- stream_session_t * s, u32 * n_trace)
+static void
+session_tx_trace_frame (vlib_main_t * vm, vlib_node_runtime_t * node,
+ u32 next_index, u32 * to_next, u16 n_segs,
+ stream_session_t * s, u32 n_trace)
{
session_queue_trace_t *t;
- vlib_trace_buffer (vm, node, next_index, b, 1 /* follow_chain */ );
- vlib_set_trace_count (vm, node, --*n_trace);
- t = vlib_add_trace (vm, node, b, sizeof (*t));
- t->session_index = s->session_index;
- t->server_thread_index = s->thread_index;
+ vlib_buffer_t *b;
+ int i;
+
+ for (i = 0; i < clib_min (n_trace, n_segs); i++)
+ {
+ b = vlib_get_buffer (vm, to_next[i - n_segs]);
+ vlib_trace_buffer (vm, node, next_index, b, 1 /* follow_chain */ );
+ t = vlib_add_trace (vm, node, b, sizeof (*t));
+ t->session_index = s->session_index;
+ t->server_thread_index = s->thread_index;
+ }
+ vlib_set_trace_count (vm, node, n_trace - i);
}
always_inline void
b->error = 0;
b->flags = VNET_BUFFER_F_LOCALLY_ORIGINATED;
b->current_data = 0;
- b->total_length_not_including_first_buffer = 0;
data0 = vlib_buffer_make_headroom (b, MAX_HDRS_LEN);
len_to_deq = clib_min (ctx->left_to_snd, ctx->deq_per_first_buf);
session_tx_fifo_chain_tail (vm, ctx, b, n_bufs, peek_data);
/* *INDENT-OFF* */
- SESSION_EVT_DBG(SESSION_EVT_DEQ, s, ({
- ed->data[0] = e->event_type;
- ed->data[1] = max_dequeue;
+ SESSION_EVT_DBG(SESSION_EVT_DEQ, ctx->s, ({
+ ed->data[0] = FIFO_EVENT_APP_TX;
+ ed->data[1] = ctx->max_dequeue;
ed->data[2] = len_to_deq;
- ed->data[3] = left_to_snd;
+ ed->data[3] = ctx->left_to_snd;
}));
/* *INDENT-ON* */
}
stream_session_t * s, int *n_tx_packets,
u8 peek_data)
{
- u32 next_index, next0, next1, next2, next3, *to_next, n_left_to_next;
+ u32 next_index, next0, next1, *to_next, n_left_to_next;
u32 n_trace = vlib_get_trace_count (vm, node), n_bufs_needed = 0;
u32 thread_index = s->thread_index, n_left, pbi;
session_manager_main_t *smm = &session_manager_main;
}
next_index = smm->session_type_to_next[s->session_type];
- next0 = next1 = next2 = next3 = next_index;
+ next0 = next1 = next_index;
tp = session_get_transport_proto (s);
ctx->s = s;
}
ctx->left_to_snd = ctx->max_len_to_snd;
n_left = ctx->n_segs_per_evt;
- while (n_left)
+
+ while (n_left >= 4)
{
- while (n_left >= 4)
- {
- vlib_buffer_t *b0, *b1;
- u32 bi0, bi1;
+ vlib_buffer_t *b0, *b1;
+ u32 bi0, bi1;
- pbi = smm->tx_buffers[thread_index][n_bufs - 3];
- pb = vlib_get_buffer (vm, pbi);
- vlib_prefetch_buffer_header (pb, STORE);
- pbi = smm->tx_buffers[thread_index][n_bufs - 4];
- pb = vlib_get_buffer (vm, pbi);
- vlib_prefetch_buffer_header (pb, STORE);
+ pbi = smm->tx_buffers[thread_index][n_bufs - 3];
+ pb = vlib_get_buffer (vm, pbi);
+ vlib_prefetch_buffer_header (pb, STORE);
+ pbi = smm->tx_buffers[thread_index][n_bufs - 4];
+ pb = vlib_get_buffer (vm, pbi);
+ vlib_prefetch_buffer_header (pb, STORE);
- to_next[0] = bi0 = smm->tx_buffers[thread_index][--n_bufs];
- to_next[1] = bi1 = smm->tx_buffers[thread_index][--n_bufs];
+ to_next[0] = bi0 = smm->tx_buffers[thread_index][--n_bufs];
+ to_next[1] = bi1 = smm->tx_buffers[thread_index][--n_bufs];
- b0 = vlib_get_buffer (vm, bi0);
- b1 = vlib_get_buffer (vm, bi1);
+ b0 = vlib_get_buffer (vm, bi0);
+ b1 = vlib_get_buffer (vm, bi1);
- session_tx_fill_buffer (vm, ctx, b0, &n_bufs, peek_data);
- session_tx_fill_buffer (vm, ctx, b1, &n_bufs, peek_data);
+ session_tx_fill_buffer (vm, ctx, b0, &n_bufs, peek_data);
+ session_tx_fill_buffer (vm, ctx, b1, &n_bufs, peek_data);
- ctx->transport_vft->push_header (ctx->tc, b0);
- ctx->transport_vft->push_header (ctx->tc, b1);
+ ctx->transport_vft->push_header (ctx->tc, b0);
+ ctx->transport_vft->push_header (ctx->tc, b1);
- to_next += 2;
- n_left_to_next -= 2;
- n_left -= 2;
+ to_next += 2;
+ n_left_to_next -= 2;
+ n_left -= 2;
- VLIB_BUFFER_TRACE_TRAJECTORY_INIT (b0);
- VLIB_BUFFER_TRACE_TRAJECTORY_INIT (b1);
+ VLIB_BUFFER_TRACE_TRAJECTORY_INIT (b0);
+ VLIB_BUFFER_TRACE_TRAJECTORY_INIT (b1);
- if (PREDICT_FALSE (n_trace > 0))
- {
- session_tx_trace_buffer (vm, node, next_index, b0, s, &n_trace);
- if (n_trace)
- session_tx_trace_buffer (vm, node, next_index, b1, s,
- &n_trace);
- }
-
- vlib_validate_buffer_enqueue_x2 (vm, node, next_index, to_next,
- n_left_to_next, bi0, bi1,
- next0, next1);
- }
- while (n_left)
- {
- vlib_buffer_t *b0;
- u32 bi0;
+ vlib_validate_buffer_enqueue_x2 (vm, node, next_index, to_next,
+ n_left_to_next, bi0, bi1, next0,
+ next1);
+ }
+ while (n_left)
+ {
+ vlib_buffer_t *b0;
+ u32 bi0;
- ASSERT (n_bufs >= 1);
- to_next[0] = bi0 = smm->tx_buffers[thread_index][--n_bufs];
- b0 = vlib_get_buffer (vm, bi0);
- session_tx_fill_buffer (vm, ctx, b0, &n_bufs, peek_data);
+ ASSERT (n_bufs >= 1);
+ to_next[0] = bi0 = smm->tx_buffers[thread_index][--n_bufs];
+ b0 = vlib_get_buffer (vm, bi0);
+ session_tx_fill_buffer (vm, ctx, b0, &n_bufs, peek_data);
- /* Ask transport to push header after current_length and
- * total_length_not_including_first_buffer are updated */
- ctx->transport_vft->push_header (ctx->tc, b0);
+ /* Ask transport to push header after current_length and
+ * total_length_not_including_first_buffer are updated */
+ ctx->transport_vft->push_header (ctx->tc, b0);
- to_next += 1;
- n_left_to_next -= 1;
- n_left -= 1;
+ to_next += 1;
+ n_left_to_next -= 1;
+ n_left -= 1;
- VLIB_BUFFER_TRACE_TRAJECTORY_INIT (b0);
- if (PREDICT_FALSE (n_trace > 0))
- session_tx_trace_buffer (vm, node, next_index, b0, s, &n_trace);
+ VLIB_BUFFER_TRACE_TRAJECTORY_INIT (b0);
- vlib_validate_buffer_enqueue_x1 (vm, node, next_index, to_next,
- n_left_to_next, bi0, next0);
- }
+ vlib_validate_buffer_enqueue_x1 (vm, node, next_index, to_next,
+ n_left_to_next, bi0, next0);
}
+
+ if (PREDICT_FALSE (n_trace > 0))
+ session_tx_trace_frame (vm, node, next_index, to_next,
+ ctx->n_segs_per_evt, s, n_trace);
+
_vec_len (smm->tx_buffers[thread_index]) = n_bufs;
*n_tx_packets += ctx->n_segs_per_evt;
vlib_put_next_frame (vm, node, next_index, n_left_to_next);
}
break;
case FIFO_EVENT_DISCONNECT:
- /* Make sure stream disconnects run after the pending list is drained */
+ /* Make sure stream disconnects run after the pending list is
+ * drained */
s0 = session_get_from_handle (e0->session_handle);
if (!e0->postponed)
{
vec_add1 (smm->pending_disconnects[thread_index], *e0);
continue;
}
- /* If tx queue is still not empty, wait a bit */
- if (svm_fifo_max_dequeue (s0->server_tx_fifo)
- && e0->postponed < 200)
+ /* If tx queue is still not empty, wait */
+ if (svm_fifo_max_dequeue (s0->server_tx_fifo))
{
- e0->postponed += 1;
vec_add1 (smm->pending_disconnects[thread_index], *e0);
continue;
}
vlib_node_increment_counter (vm, session_queue_node.index,
SESSION_QUEUE_ERROR_TX, n_tx_packets);
- SESSION_EVT_DBG (SESSION_EVT_DEQ_NODE, 1);
+ SESSION_EVT_DBG (SESSION_EVT_DISPATCH_END, smm, thread_index);
return n_tx_packets;
}