vlib_buffer_get_trace_index (b)));
}
-void vlib_add_handoff_trace (vlib_main_t * vm, vlib_buffer_t * b);
+int vlib_add_handoff_trace (vlib_main_t * vm, vlib_buffer_t * b);
always_inline void *
vlib_add_trace_inline (vlib_main_t * vm,
/* Are we trying to trace a handoff case? */
if (PREDICT_FALSE (vlib_buffer_get_trace_thread (b) != vm->thread_index))
- vlib_add_handoff_trace (vm, b);
+ if (PREDICT_FALSE (!vlib_add_handoff_trace (vm, b)))
+ return vnet_trace_placeholder;
vlib_validate_trace (tm, b);
u32 classify_table_index, int func);
-/* Mark buffer as traced and allocate trace buffer. */
-always_inline void
+/*
+ * Mark buffer as traced and allocate trace buffer.
+ * return 1 if the buffer is successfully traced, 0 if not
+ * A buffer might not be traced if tracing is off or if the packet did not
+ * match the filter.
+ */
+always_inline __clib_warn_unused_result int
vlib_trace_buffer (vlib_main_t * vm,
vlib_node_runtime_t * r,
u32 next_index, vlib_buffer_t * b, int follow_chain)
vlib_trace_header_t **h;
if (PREDICT_FALSE (tm->trace_enable == 0))
- return;
+ return 0;
/* Classifier filter in use? */
if (PREDICT_FALSE (vlib_global_main.trace_filter.trace_filter_enable))
if (vnet_is_packet_traced
(b, vlib_global_main.trace_filter.trace_classify_table_index,
0 /* full classify */ ) != 1)
- return;
+ return 0;
}
/*
(vm->thread_index, h - tm->trace_buffer_pool);
}
while (follow_chain && (b = vlib_get_next_buffer (vm, b)));
+
+ return 1;
}
always_inline void
{
vlib_trace_main_t *tm = &vm->trace_main;
vlib_trace_node_t *tn;
- int n;
if (rt->node_index >= vec_len (tm->nodes))
return 0;
tn = tm->nodes + rt->node_index;
- n = tn->limit - tn->count;
- ASSERT (n >= 0);
+ ASSERT (tn->count <= tn->limit);
- return n;
+ return tn->limit - tn->count;
}
always_inline void