a->sep.sw_if_index = ENDPOINT_INVALID_INDEX;
a->sep.transport_proto = mp->proto;
a->sep_ext.ckpair_index = mp->ckpair_index;
+ a->sep_ext.crypto_engine = mp->crypto_engine;
a->app_index = app->app_index;
a->wrk_map_index = mp->wrk_index;
a->sep.peer.sw_if_index = ENDPOINT_INVALID_INDEX;
a->sep_ext.parent_handle = mp->parent_handle;
a->sep_ext.ckpair_index = mp->ckpair_index;
+ a->sep_ext.crypto_engine = mp->crypto_engine;
if (mp->hostname_len)
{
vec_validate (a->sep_ext.hostname, mp->hostname_len - 1);
}
ctx->snd_mss = ctx->transport_vft->send_mss (ctx->tc);
- ctx->snd_space = transport_connection_snd_space (ctx->tc,
- vm->clib_time.
- last_cpu_time,
- ctx->snd_mss);
-
- if (ctx->snd_space == 0 || ctx->snd_mss == 0)
+ if (PREDICT_FALSE (ctx->snd_mss == 0))
{
session_evt_add_old (wrk, elt);
return SESSION_TX_NO_DATA;
}
+ ctx->snd_space = transport_connection_snd_space (ctx->tc);
+
+ /* This flow queue is "empty" so it should be re-evaluated before
+ * the ones that have data to send. */
+ if (!ctx->snd_space)
+ {
+ session_evt_add_head_old (wrk, elt);
+ return SESSION_TX_NO_DATA;
+ }
+
+ if (transport_connection_is_tx_paced (ctx->tc))
+ {
+ u32 snd_space = transport_connection_tx_pacer_burst (ctx->tc);
+ if (snd_space < TRANSPORT_PACER_MIN_BURST)
+ {
+ session_evt_add_head_old (wrk, elt);
+ return SESSION_TX_NO_DATA;
+ }
+ snd_space = clib_min (ctx->snd_space, snd_space);
+ ctx->snd_space = snd_space >= ctx->snd_mss ?
+ snd_space - snd_space % ctx->snd_mss : snd_space;
+ }
+
/* Allow enqueuing of a new event */
svm_fifo_unset_event (ctx->s->tx_fifo);
if (PREDICT_FALSE (!ctx->max_len_to_snd))
{
- transport_connection_tx_pacer_reset_bucket (ctx->tc,
- vm->clib_time.
- last_cpu_time);
+ transport_connection_tx_pacer_reset_bucket (ctx->tc, 0);
return SESSION_TX_NO_DATA;
}
{
if (n_bufs)
vlib_buffer_free (vm, wrk->tx_buffers, n_bufs);
- session_evt_add_old (wrk, elt);
+ if (svm_fifo_set_event (ctx->s->tx_fifo))
+ session_evt_add_head_old (wrk, elt);
vlib_node_increment_counter (wrk->vm, node->node_index,
SESSION_QUEUE_ERROR_NO_BUFFER, 1);
return SESSION_TX_NO_BUFFERS;
SESSION_EVT (SESSION_EVT_DISPATCH_START, wrk);
wrk->last_vlib_time = vlib_time_now (vm);
+ wrk->last_vlib_us_time = wrk->last_vlib_time * CLIB_US_TIME_FREQ;
/*
* Update transport time
if (old_ti != wrk->old_head)
{
+ clib_llist_index_t ei, next_ei;
+
old_he = pool_elt_at_index (wrk->event_elts, wrk->old_head);
+ ei = clib_llist_next_index (old_he, evt_list);
+
while (n_tx_packets < VLIB_FRAME_SIZE)
{
- clib_llist_index_t ei;
+ elt = pool_elt_at_index (wrk->event_elts, ei);
+ next_ei = clib_llist_next_index (elt, evt_list);
+ clib_llist_remove (wrk->event_elts, evt_list, elt);
- clib_llist_pop_first (wrk->event_elts, evt_list, elt, old_he);
- ei = clib_llist_entry_index (wrk->event_elts, elt);
session_event_dispatch_io (wrk, node, elt, thread_index,
&n_tx_packets);
- old_he = pool_elt_at_index (wrk->event_elts, wrk->old_head);
if (ei == old_ti)
break;
+
+ ei = next_ei;
};
}