ASSERT (vnet_trace_dummy);
+ if (PREDICT_FALSE ((b->flags & VLIB_BUFFER_IS_TRACED) == 0))
+ return vnet_trace_dummy;
+
if (PREDICT_FALSE (tm->add_trace_callback != 0))
{
return tm->add_trace_callback ((struct vlib_main_t *) vm,
}
void trace_apply_filter (vlib_main_t * vm);
+int vnet_is_packet_traced (vlib_buffer_t * b,
+ u32 classify_table_index, int func);
+
/* Mark buffer as traced and allocate trace buffer. */
always_inline void
if (PREDICT_FALSE (tm->trace_enable == 0))
return;
+ /* Classifier filter in use? */
+ if (PREDICT_FALSE (vlib_global_main.trace_filter.trace_filter_enable))
+ {
+ /* See if we're supposed to trace this packet... */
+ if (vnet_is_packet_traced
+ (b, vlib_global_main.trace_filter.trace_classify_table_index,
+ 0 /* full classify */ ) != 1)
+ return;
+ }
+
/*
* Apply filter to existing traces to keep number of allocated traces low.
* Performed each time around the main loop.
(struct vlib_trace_main_t *) tm);
}
- //vlib_trace_next_frame (vm, r, next_index);
+ vlib_trace_next_frame (vm, r, next_index);
pool_get (tm->trace_buffer_pool, h);