{
if (cf)
{
- vec_validate (ct->buffer_indice, n_cache + cf->n_elts);
+ vec_validate (ct->buffer_indices, n_cache + cf->n_elts);
vec_validate (ct->nexts, n_cache + cf->n_elts);
- clib_memcpy_fast (ct->buffer_indice + n_cache, cf->buffer_indices,
+ clib_memcpy_fast (ct->buffer_indices + n_cache, cf->buffer_indices,
sizeof (u32) * cf->n_elts);
if (cf->state == VNET_CRYPTO_FRAME_STATE_SUCCESS)
{
n_cache += cf->n_elts;
if (n_cache >= VLIB_FRAME_SIZE)
{
- vlib_buffer_enqueue_to_next (vm, node, ct->buffer_indice,
+ vlib_buffer_enqueue_to_next (vm, node, ct->buffer_indices,
ct->nexts, n_cache);
n_cache = 0;
}
if (cm->dispatch_mode == VNET_CRYPTO_ASYNC_DISPATCH_INTERRUPT
&& n_elts > 0)
{
- vlib_node_set_interrupt_pending (vlib_mains[enqueue_thread_idx],
- cm->crypto_node_index);
+ vlib_node_set_interrupt_pending (
+ vlib_get_main_by_index (enqueue_thread_idx),
+ cm->crypto_node_index);
}
n_elts = 0;
u32 index;
/* *INDENT-OFF* */
- clib_bitmap_foreach (index, cm->async_active_ids, ({
+ clib_bitmap_foreach (index, cm->async_active_ids) {
n_cache = crypto_dequeue_frame (vm, node, ct, cm->dequeue_handlers[index],
n_cache, &n_dispatched);
- }));
+ }
/* *INDENT-ON* */
if (n_cache)
- vlib_buffer_enqueue_to_next (vm, node, ct->buffer_indice, ct->nexts,
+ vlib_buffer_enqueue_to_next (vm, node, ct->buffer_indices, ct->nexts,
n_cache);
return n_dispatched;