/* do worker handoff based on thread_index in NAT HA protcol header */
static_always_inline uword
-ipsec_handoff (vlib_main_t * vm,
- vlib_node_runtime_t * node,
- vlib_frame_t * frame, u32 fq_index, bool is_enc)
+ipsec_handoff (vlib_main_t *vm, vlib_node_runtime_t *node, vlib_frame_t *frame,
+ u32 fq_index)
{
vlib_buffer_t *bufs[VLIB_FRAME_SIZE], **b;
u16 thread_indices[VLIB_FRAME_SIZE], *ti;
u32 n_enq, n_left_from, *from;
- ipsec_main_t *im;
- im = &ipsec_main;
from = vlib_frame_vector_args (frame);
n_left_from = frame->n_vectors;
vlib_get_buffers (vm, from, bufs, n_left_from);
while (n_left_from >= 4)
{
- ipsec_sa_t *sa0, *sa1, *sa2, *sa3;
- u32 sai0, sai1, sai2, sai3;
-
/* Prefetch next iteration. */
if (n_left_from >= 12)
{
vlib_prefetch_buffer_header (b[9], LOAD);
vlib_prefetch_buffer_header (b[10], LOAD);
vlib_prefetch_buffer_header (b[11], LOAD);
-
- vlib_prefetch_buffer_data (b[4], LOAD);
- vlib_prefetch_buffer_data (b[5], LOAD);
- vlib_prefetch_buffer_data (b[6], LOAD);
- vlib_prefetch_buffer_data (b[7], LOAD);
}
- sai0 = vnet_buffer (b[0])->ipsec.sad_index;
- sai1 = vnet_buffer (b[1])->ipsec.sad_index;
- sai2 = vnet_buffer (b[2])->ipsec.sad_index;
- sai3 = vnet_buffer (b[3])->ipsec.sad_index;
- sa0 = pool_elt_at_index (im->sad, sai0);
- sa1 = pool_elt_at_index (im->sad, sai1);
- sa2 = pool_elt_at_index (im->sad, sai2);
- sa3 = pool_elt_at_index (im->sad, sai3);
+ ti[0] = vnet_buffer (b[0])->ipsec.thread_index;
+ ti[1] = vnet_buffer (b[1])->ipsec.thread_index;
+ ti[2] = vnet_buffer (b[2])->ipsec.thread_index;
+ ti[3] = vnet_buffer (b[3])->ipsec.thread_index;
- if (is_enc)
- {
- ti[0] = sa0->encrypt_thread_index;
- ti[1] = sa1->encrypt_thread_index;
- ti[2] = sa2->encrypt_thread_index;
- ti[3] = sa3->encrypt_thread_index;
- }
- else
- {
- ti[0] = sa0->decrypt_thread_index;
- ti[1] = sa1->decrypt_thread_index;
- ti[2] = sa2->decrypt_thread_index;
- ti[3] = sa3->decrypt_thread_index;
- }
-
- if (node->flags & VLIB_NODE_FLAG_TRACE)
+ if (PREDICT_FALSE (node->flags & VLIB_NODE_FLAG_TRACE))
{
if (PREDICT_FALSE (b[0]->flags & VLIB_BUFFER_IS_TRACED))
{
}
while (n_left_from > 0)
{
- ipsec_sa_t *sa0;
- u32 sai0;
-
- sai0 = vnet_buffer (b[0])->ipsec.sad_index;
- sa0 = pool_elt_at_index (im->sad, sai0);
-
- if (is_enc)
- ti[0] = sa0->encrypt_thread_index;
- else
- ti[0] = sa0->decrypt_thread_index;
+ ti[0] = vnet_buffer (b[0])->ipsec.thread_index;
if (PREDICT_FALSE (b[0]->flags & VLIB_BUFFER_IS_TRACED))
{
b += 1;
}
- n_enq = vlib_buffer_enqueue_to_thread (vm, fq_index, from,
+ n_enq = vlib_buffer_enqueue_to_thread (vm, node, fq_index, from,
thread_indices, frame->n_vectors, 1);
if (n_enq < frame->n_vectors)
{
ipsec_main_t *im = &ipsec_main;
- return ipsec_handoff (vm, node, from_frame, im->esp4_enc_fq_index, true);
+ return ipsec_handoff (vm, node, from_frame, im->esp4_enc_fq_index);
}
VLIB_NODE_FN (esp6_encrypt_handoff) (vlib_main_t * vm,
{
ipsec_main_t *im = &ipsec_main;
- return ipsec_handoff (vm, node, from_frame, im->esp6_enc_fq_index, true);
+ return ipsec_handoff (vm, node, from_frame, im->esp6_enc_fq_index);
}
VLIB_NODE_FN (esp4_encrypt_tun_handoff) (vlib_main_t * vm,
{
ipsec_main_t *im = &ipsec_main;
- return ipsec_handoff (vm, node, from_frame, im->esp4_enc_tun_fq_index,
- true);
+ return ipsec_handoff (vm, node, from_frame, im->esp4_enc_tun_fq_index);
}
VLIB_NODE_FN (esp6_encrypt_tun_handoff) (vlib_main_t * vm,
{
ipsec_main_t *im = &ipsec_main;
- return ipsec_handoff (vm, node, from_frame, im->esp6_enc_tun_fq_index,
- true);
+ return ipsec_handoff (vm, node, from_frame, im->esp6_enc_tun_fq_index);
}
VLIB_NODE_FN (esp_mpls_encrypt_tun_handoff)
{
ipsec_main_t *im = &ipsec_main;
- return ipsec_handoff (vm, node, from_frame, im->esp_mpls_enc_tun_fq_index,
- true);
+ return ipsec_handoff (vm, node, from_frame, im->esp_mpls_enc_tun_fq_index);
}
VLIB_NODE_FN (esp4_decrypt_handoff) (vlib_main_t * vm,
{
ipsec_main_t *im = &ipsec_main;
- return ipsec_handoff (vm, node, from_frame, im->esp4_dec_fq_index, false);
+ return ipsec_handoff (vm, node, from_frame, im->esp4_dec_fq_index);
}
VLIB_NODE_FN (esp6_decrypt_handoff) (vlib_main_t * vm,
{
ipsec_main_t *im = &ipsec_main;
- return ipsec_handoff (vm, node, from_frame, im->esp6_dec_fq_index, false);
+ return ipsec_handoff (vm, node, from_frame, im->esp6_dec_fq_index);
}
VLIB_NODE_FN (esp4_decrypt_tun_handoff) (vlib_main_t * vm,
{
ipsec_main_t *im = &ipsec_main;
- return ipsec_handoff (vm, node, from_frame, im->esp4_dec_tun_fq_index,
- false);
+ return ipsec_handoff (vm, node, from_frame, im->esp4_dec_tun_fq_index);
}
VLIB_NODE_FN (esp6_decrypt_tun_handoff) (vlib_main_t * vm,
{
ipsec_main_t *im = &ipsec_main;
- return ipsec_handoff (vm, node, from_frame, im->esp6_dec_tun_fq_index,
- false);
+ return ipsec_handoff (vm, node, from_frame, im->esp6_dec_tun_fq_index);
}
VLIB_NODE_FN (ah4_encrypt_handoff) (vlib_main_t * vm,
{
ipsec_main_t *im = &ipsec_main;
- return ipsec_handoff (vm, node, from_frame, im->ah4_enc_fq_index, true);
+ return ipsec_handoff (vm, node, from_frame, im->ah4_enc_fq_index);
}
VLIB_NODE_FN (ah6_encrypt_handoff) (vlib_main_t * vm,
{
ipsec_main_t *im = &ipsec_main;
- return ipsec_handoff (vm, node, from_frame, im->ah6_enc_fq_index, true);
+ return ipsec_handoff (vm, node, from_frame, im->ah6_enc_fq_index);
}
VLIB_NODE_FN (ah4_decrypt_handoff) (vlib_main_t * vm,
{
ipsec_main_t *im = &ipsec_main;
- return ipsec_handoff (vm, node, from_frame, im->ah4_dec_fq_index, false);
+ return ipsec_handoff (vm, node, from_frame, im->ah4_dec_fq_index);
}
VLIB_NODE_FN (ah6_decrypt_handoff) (vlib_main_t * vm,
{
ipsec_main_t *im = &ipsec_main;
- return ipsec_handoff (vm, node, from_frame, im->ah6_dec_fq_index, false);
+ return ipsec_handoff (vm, node, from_frame, im->ah6_dec_fq_index);
}
-/* *INDENT-OFF* */
VLIB_REGISTER_NODE (esp4_encrypt_handoff) = {
.name = "esp4-encrypt-handoff",
.vector_size = sizeof (u32),
[0] = "error-drop",
},
};
-/* *INDENT-ON* */
/*
* fd.io coding-style-patch-verification: ON