a->sep.sw_if_index = ENDPOINT_INVALID_INDEX;
a->sep.transport_proto = mp->proto;
a->sep_ext.ckpair_index = mp->ckpair_index;
+ a->sep_ext.crypto_engine = mp->crypto_engine;
a->app_index = app->app_index;
a->wrk_map_index = mp->wrk_index;
a->sep.port = mp->port;
a->sep.transport_proto = mp->proto;
a->sep.peer.fib_index = mp->vrf;
+ clib_memcpy_fast (&a->sep.peer.ip, &mp->lcl_ip, sizeof (mp->lcl_ip));
a->sep.peer.sw_if_index = ENDPOINT_INVALID_INDEX;
a->sep_ext.parent_handle = mp->parent_handle;
a->sep_ext.ckpair_index = mp->ckpair_index;
+ a->sep_ext.crypto_engine = mp->crypto_engine;
if (mp->hostname_len)
{
vec_validate (a->sep_ext.hostname, mp->hostname_len - 1);
}
ctx->snd_mss = ctx->transport_vft->send_mss (ctx->tc);
- ctx->snd_space = transport_connection_snd_space (ctx->tc,
- vm->clib_time.
- last_cpu_time,
- ctx->snd_mss);
-
- if (ctx->snd_space == 0 || ctx->snd_mss == 0)
+ if (PREDICT_FALSE (ctx->snd_mss == 0))
{
session_evt_add_old (wrk, elt);
return SESSION_TX_NO_DATA;
}
+ ctx->snd_space = transport_connection_snd_space (ctx->tc, ctx->snd_mss);
+
+ /* This flow queue is "empty" so it should be re-evaluated before
+ * the ones that have data to send. */
+ if (ctx->snd_space == 0)
+ {
+ session_evt_add_head_old (wrk, elt);
+ return SESSION_TX_NO_DATA;
+ }
+
/* Allow enqueuing of a new event */
svm_fifo_unset_event (ctx->s->tx_fifo);
if (PREDICT_FALSE (!ctx->max_len_to_snd))
{
- transport_connection_tx_pacer_reset_bucket (ctx->tc,
- vm->clib_time.
- last_cpu_time);
+ transport_connection_tx_pacer_reset_bucket (ctx->tc);
return SESSION_TX_NO_DATA;
}
{
if (n_bufs)
vlib_buffer_free (vm, wrk->tx_buffers, n_bufs);
- session_evt_add_old (wrk, elt);
+ if (svm_fifo_set_event (ctx->s->tx_fifo))
+ session_evt_add_head_old (wrk, elt);
vlib_node_increment_counter (wrk->vm, node->node_index,
SESSION_QUEUE_ERROR_NO_BUFFER, 1);
return SESSION_TX_NO_BUFFERS;
SESSION_EVT (SESSION_EVT_DISPATCH_START, wrk);
wrk->last_vlib_time = vlib_time_now (vm);
+ wrk->last_vlib_us_time = wrk->last_vlib_time * CLIB_US_TIME_FREQ;
/*
* Update transport time
*/
new_he = pool_elt_at_index (wrk->event_elts, wrk->new_head);
+ old_he = pool_elt_at_index (wrk->event_elts, wrk->old_head);
+ old_ti = clib_llist_prev_index (old_he, evt_list);
/* *INDENT-OFF* */
clib_llist_foreach_safe (wrk->event_elts, evt_list, new_he, elt, ({
/* *INDENT-ON* */
/*
- * Handle the old io events
+ * Handle the old io events, if we had any prior to processing the new ones
*/
- old_he = pool_elt_at_index (wrk->event_elts, wrk->old_head);
- old_ti = clib_llist_prev_index (old_he, evt_list);
-
- while (n_tx_packets < VLIB_FRAME_SIZE
- && !clib_llist_is_empty (wrk->event_elts, evt_list, old_he))
+ if (old_ti != wrk->old_head)
{
- clib_llist_index_t ei;
-
- clib_llist_pop_first (wrk->event_elts, evt_list, elt, old_he);
- ei = clib_llist_entry_index (wrk->event_elts, elt);
- session_event_dispatch_io (wrk, node, elt, thread_index, &n_tx_packets);
+ clib_llist_index_t ei, next_ei;
old_he = pool_elt_at_index (wrk->event_elts, wrk->old_head);
- if (ei == old_ti)
- break;
- };
+ ei = clib_llist_next_index (old_he, evt_list);
+
+ while (n_tx_packets < VLIB_FRAME_SIZE)
+ {
+ elt = pool_elt_at_index (wrk->event_elts, ei);
+ next_ei = clib_llist_next_index (elt, evt_list);
+ clib_llist_remove (wrk->event_elts, evt_list, elt);
+
+ session_event_dispatch_io (wrk, node, elt, thread_index,
+ &n_tx_packets);
+
+ if (ei == old_ti)
+ break;
+
+ ei = next_ei;
+ };
+ }
if (vec_len (wrk->pending_tx_buffers))
session_flush_pending_tx_buffers (wrk, node);