return s;
}
-vlib_node_registration_t nat64_in2out_node;
-vlib_node_registration_t nat64_in2out_slowpath_node;
-vlib_node_registration_t nat64_in2out_reass_node;
-vlib_node_registration_t nat64_in2out_handoff_node;
#define foreach_nat64_in2out_error \
_(UNSUPPORTED_PROTOCOL, "unsupported protocol") \
u32 pkts_processed = 0;
u32 stats_node_index;
u32 thread_index = vm->thread_index;
+ nat64_main_t *nm = &nat64_main;
+
u32 tcp_packets = 0, udp_packets = 0, icmp_packets = 0, other_packets =
0, fragments = 0;
stats_node_index =
- is_slow_path ? nat64_in2out_slowpath_node.index : nat64_in2out_node.index;
+ is_slow_path ? nm->in2out_slowpath_node_index : nm->in2out_node_index;
from = vlib_frame_vector_args (frame);
n_left_from = frame->n_vectors;
return frame->n_vectors;
}
-static uword
-nat64_in2out_node_fn (vlib_main_t * vm, vlib_node_runtime_t * node,
- vlib_frame_t * frame)
+VLIB_NODE_FN (nat64_in2out_node) (vlib_main_t * vm,
+ vlib_node_runtime_t * node,
+ vlib_frame_t * frame)
{
return nat64_in2out_node_fn_inline (vm, node, frame, 0);
}
/* *INDENT-OFF* */
VLIB_REGISTER_NODE (nat64_in2out_node) = {
- .function = nat64_in2out_node_fn,
.name = "nat64-in2out",
.vector_size = sizeof (u32),
.format_trace = format_nat64_in2out_trace,
};
/* *INDENT-ON* */
-VLIB_NODE_FUNCTION_MULTIARCH (nat64_in2out_node, nat64_in2out_node_fn);
-
-static uword
-nat64_in2out_slowpath_node_fn (vlib_main_t * vm, vlib_node_runtime_t * node,
- vlib_frame_t * frame)
+VLIB_NODE_FN (nat64_in2out_slowpath_node) (vlib_main_t * vm,
+ vlib_node_runtime_t * node,
+ vlib_frame_t * frame)
{
return nat64_in2out_node_fn_inline (vm, node, frame, 1);
}
/* *INDENT-OFF* */
VLIB_REGISTER_NODE (nat64_in2out_slowpath_node) = {
- .function = nat64_in2out_slowpath_node_fn,
.name = "nat64-in2out-slowpath",
.vector_size = sizeof (u32),
.format_trace = format_nat64_in2out_trace,
};
/* *INDENT-ON* */
-VLIB_NODE_FUNCTION_MULTIARCH (nat64_in2out_slowpath_node,
- nat64_in2out_slowpath_node_fn);
-
typedef struct nat64_in2out_frag_set_ctx_t_
{
vlib_main_t *vm;
return 0;
}
-static uword
-nat64_in2out_reass_node_fn (vlib_main_t * vm,
- vlib_node_runtime_t * node, vlib_frame_t * frame)
+VLIB_NODE_FN (nat64_in2out_reass_node) (vlib_main_t * vm,
+ vlib_node_runtime_t * node,
+ vlib_frame_t * frame)
{
u32 n_left_from, *from, *to_next;
nat64_in2out_next_t next_index;
vlib_put_next_frame (vm, node, next_index, n_left_to_next);
}
- vlib_node_increment_counter (vm, nat64_in2out_reass_node.index,
+ vlib_node_increment_counter (vm, nm->in2out_reass_node_index,
NAT64_IN2OUT_ERROR_PROCESSED_FRAGMENTS,
pkts_processed);
- vlib_node_increment_counter (vm, nat64_in2out_reass_node.index,
+ vlib_node_increment_counter (vm, nm->in2out_reass_node_index,
NAT64_IN2OUT_ERROR_CACHED_FRAGMENTS,
cached_fragments);
/* *INDENT-OFF* */
VLIB_REGISTER_NODE (nat64_in2out_reass_node) = {
- .function = nat64_in2out_reass_node_fn,
.name = "nat64-in2out-reass",
.vector_size = sizeof (u32),
.format_trace = format_nat64_in2out_reass_trace,
};
/* *INDENT-ON* */
-VLIB_NODE_FUNCTION_MULTIARCH (nat64_in2out_reass_node,
- nat64_in2out_reass_node_fn);
-
#define foreach_nat64_in2out_handoff_error \
_(CONGESTION_DROP, "congestion drop") \
_(SAME_WORKER, "same worker") \
return s;
}
-static inline uword
-nat64_in2out_handoff_node_fn (vlib_main_t * vm, vlib_node_runtime_t * node,
- vlib_frame_t * frame)
+VLIB_NODE_FN (nat64_in2out_handoff_node) (vlib_main_t * vm,
+ vlib_node_runtime_t * node,
+ vlib_frame_t * frame)
{
nat64_main_t *nm = &nat64_main;
vlib_buffer_t *bufs[VLIB_FRAME_SIZE], **b;
/* *INDENT-OFF* */
VLIB_REGISTER_NODE (nat64_in2out_handoff_node) = {
- .function = nat64_in2out_handoff_node_fn,
.name = "nat64-in2out-handoff",
.vector_size = sizeof (u32),
.format_trace = format_nat64_in2out_handoff_trace,
};
/* *INDENT-ON* */
-VLIB_NODE_FUNCTION_MULTIARCH (nat64_in2out_handoff_node,
- nat64_in2out_handoff_node_fn);
-
/*
* fd.io coding-style-patch-verification: ON
*