X-Git-Url: https://gerrit.fd.io/r/gitweb?a=blobdiff_plain;f=src%2Fvlib%2Fmain.c;h=4223474367ec549182186ac6243eaf95d891def0;hb=d7b306657;hp=b46caf1bcb444817342230093849cbc8e7bf3e5a;hpb=a3d5986a35743a3ca1020155bed8267b5246d155;p=vpp.git diff --git a/src/vlib/main.c b/src/vlib/main.c index b46caf1bcb4..4223474367e 100644 --- a/src/vlib/main.c +++ b/src/vlib/main.c @@ -52,8 +52,6 @@ CJ_GLOBAL_LOG_PROTOTYPE; speculative vector enqueues which overflow vector data in next frame. */ #define VLIB_FRAME_SIZE_ALLOC (VLIB_FRAME_SIZE + 4) -u32 wraps; - always_inline u32 vlib_frame_bytes (u32 n_scalar_bytes, u32 n_vector_bytes) { @@ -91,10 +89,11 @@ vlib_frame_find_magic (vlib_frame_t * f, vlib_node_t * node) return p; } -static vlib_frame_size_t * +static inline vlib_frame_size_t * get_frame_size_info (vlib_node_main_t * nm, u32 n_scalar_bytes, u32 n_vector_bytes) { +#ifdef VLIB_SUPPORTS_ARBITRARY_SCALAR_SIZES uword key = (n_scalar_bytes << 16) | n_vector_bytes; uword *p, i; @@ -109,9 +108,14 @@ get_frame_size_info (vlib_node_main_t * nm, } return vec_elt_at_index (nm->frame_sizes, i); +#else + ASSERT (vlib_frame_bytes (n_scalar_bytes, n_vector_bytes) + == (vlib_frame_bytes (0, 4))); + return vec_elt_at_index (nm->frame_sizes, 0); +#endif } -static u32 +static vlib_frame_t * vlib_frame_alloc_to_node (vlib_main_t * vm, u32 to_node_index, u32 frame_flags) { @@ -119,7 +123,7 @@ vlib_frame_alloc_to_node (vlib_main_t * vm, u32 to_node_index, vlib_frame_size_t *fs; vlib_node_t *to_node; vlib_frame_t *f; - u32 fi, l, n, scalar_size, vector_size; + u32 l, n, scalar_size, vector_size; to_node = vlib_get_node (vm, to_node_index); @@ -128,17 +132,15 @@ vlib_frame_alloc_to_node (vlib_main_t * vm, u32 to_node_index, fs = get_frame_size_info (nm, scalar_size, vector_size); n = vlib_frame_bytes (scalar_size, vector_size); - if ((l = vec_len (fs->free_frame_indices)) > 0) + if ((l = vec_len (fs->free_frames)) > 0) { /* Allocate from end of free list. */ - fi = fs->free_frame_indices[l - 1]; - f = vlib_get_frame_no_check (vm, fi); - _vec_len (fs->free_frame_indices) = l - 1; + f = fs->free_frames[l - 1]; + _vec_len (fs->free_frames) = l - 1; } else { f = clib_mem_alloc_aligned_no_fail (n, VLIB_FRAME_ALIGN); - fi = vlib_frame_index_no_check (vm, f); } /* Poison frame when debugging. */ @@ -161,12 +163,12 @@ vlib_frame_alloc_to_node (vlib_main_t * vm, u32 to_node_index, fs->n_alloc_frames += 1; - return fi; + return f; } /* Allocate a frame for from FROM_NODE to TO_NODE via TO_NEXT_INDEX. Returns frame index. */ -static u32 +static vlib_frame_t * vlib_frame_alloc (vlib_main_t * vm, vlib_node_runtime_t * from_node_runtime, u32 to_next_index) { @@ -182,10 +184,10 @@ vlib_frame_alloc (vlib_main_t * vm, vlib_node_runtime_t * from_node_runtime, vlib_frame_t * vlib_get_frame_to_node (vlib_main_t * vm, u32 to_node_index) { - u32 fi = vlib_frame_alloc_to_node (vm, to_node_index, - /* frame_flags */ - VLIB_FRAME_FREE_AFTER_DISPATCH); - return vlib_get_frame (vm, fi); + vlib_frame_t *f = vlib_frame_alloc_to_node (vm, to_node_index, + /* frame_flags */ + VLIB_FRAME_FREE_AFTER_DISPATCH); + return vlib_get_frame (vm, f); } void @@ -202,7 +204,7 @@ vlib_put_frame_to_node (vlib_main_t * vm, u32 to_node_index, vlib_frame_t * f) vec_add2 (vm->node_main.pending_frames, p, 1); f->frame_flags |= VLIB_FRAME_PENDING; - p->frame_index = vlib_frame_index (vm, f); + p->frame = vlib_get_frame (vm, f); p->node_runtime_index = to_node->runtime_index; p->next_frame_index = VLIB_PENDING_FRAME_NO_NEXT_FRAME; } @@ -214,28 +216,24 @@ vlib_frame_free (vlib_main_t * vm, vlib_node_runtime_t * r, vlib_frame_t * f) vlib_node_main_t *nm = &vm->node_main; vlib_node_t *node; vlib_frame_size_t *fs; - u32 frame_index; ASSERT (f->frame_flags & VLIB_FRAME_IS_ALLOCATED); node = vlib_get_node (vm, r->node_index); fs = get_frame_size_info (nm, node->scalar_size, node->vector_size); - frame_index = vlib_frame_index (vm, f); - ASSERT (f->frame_flags & VLIB_FRAME_IS_ALLOCATED); /* No next frames may point to freed frame. */ if (CLIB_DEBUG > 0) { vlib_next_frame_t *nf; - vec_foreach (nf, vm->node_main.next_frames) - ASSERT (nf->frame_index != frame_index); + vec_foreach (nf, vm->node_main.next_frames) ASSERT (nf->frame != f); } - f->frame_flags &= ~VLIB_FRAME_IS_ALLOCATED; + f->frame_flags &= ~(VLIB_FRAME_IS_ALLOCATED | VLIB_FRAME_NO_APPEND); - vec_add1 (fs->free_frame_indices, frame_index); + vec_add1 (fs->free_frames, f); ASSERT (fs->n_alloc_frames > 0); fs->n_alloc_frames -= 1; } @@ -251,7 +249,7 @@ show_frame_stats (vlib_main_t * vm, vec_foreach (fs, nm->frame_sizes) { u32 n_alloc = fs->n_alloc_frames; - u32 n_free = vec_len (fs->free_frame_indices); + u32 n_free = vec_len (fs->free_frames); if (n_alloc + n_free > 0) vlib_cli_output (vm, "%=6d%=12d%=12d", @@ -315,11 +313,11 @@ vlib_next_frame_change_ownership (vlib_main_t * vm, if (next_frame->flags & VLIB_FRAME_PENDING) { vlib_pending_frame_t *p; - if (next_frame->frame_index != ~0) + if (next_frame->frame != NULL) { vec_foreach (p, nm->pending_frames) { - if (p->frame_index == next_frame->frame_index) + if (p->frame == next_frame->frame) { p->next_frame_index = next_frame - vm->node_main.next_frames; @@ -371,11 +369,11 @@ vlib_get_next_frame_internal (vlib_main_t * vm, /* ??? Don't need valid flag: can use frame_index == ~0 */ if (PREDICT_FALSE (!(nf->flags & VLIB_FRAME_IS_ALLOCATED))) { - nf->frame_index = vlib_frame_alloc (vm, node, next_index); + nf->frame = vlib_frame_alloc (vm, node, next_index); nf->flags |= VLIB_FRAME_IS_ALLOCATED; } - f = vlib_get_frame (vm, nf->frame_index); + f = nf->frame; /* Has frame been removed from pending vector (e.g. finished dispatching)? If so we can reuse frame. */ @@ -384,23 +382,25 @@ vlib_get_next_frame_internal (vlib_main_t * vm, { nf->flags &= ~VLIB_FRAME_PENDING; f->n_vectors = 0; + f->flags = 0; } - /* Allocate new frame if current one is already full. */ + /* Allocate new frame if current one is marked as no-append or + it is already full. */ n_used = f->n_vectors; - if (n_used >= VLIB_FRAME_SIZE || (allocate_new_next_frame && n_used > 0)) + if (n_used >= VLIB_FRAME_SIZE || (allocate_new_next_frame && n_used > 0) || + (f->frame_flags & VLIB_FRAME_NO_APPEND)) { /* Old frame may need to be freed after dispatch, since we'll have two redundant frames from node -> next node. */ if (!(nf->flags & VLIB_FRAME_NO_FREE_AFTER_DISPATCH)) { - vlib_frame_t *f_old = vlib_get_frame (vm, nf->frame_index); + vlib_frame_t *f_old = vlib_get_frame (vm, nf->frame); f_old->frame_flags |= VLIB_FRAME_FREE_AFTER_DISPATCH; } /* Allocate new frame to replace full one. */ - nf->frame_index = vlib_frame_alloc (vm, node, next_index); - f = vlib_get_frame (vm, nf->frame_index); + f = nf->frame = vlib_frame_alloc (vm, node, next_index); n_used = f->n_vectors; } @@ -429,7 +429,7 @@ vlib_put_next_frame_validate (vlib_main_t * vm, u32 n_before, n_after; nf = vlib_node_runtime_get_next_frame (vm, rt, next_index); - f = vlib_get_frame (vm, nf->frame_index); + f = vlib_get_frame (vm, nf->frame); ASSERT (n_vectors_left <= VLIB_FRAME_SIZE); n_after = VLIB_FRAME_SIZE - n_vectors_left; @@ -462,11 +462,11 @@ vlib_put_next_frame (vlib_main_t * vm, vlib_frame_t *f; u32 n_vectors_in_frame; - if (buffer_main.callbacks_registered == 0 && CLIB_DEBUG > 0) + if (CLIB_DEBUG > 0) vlib_put_next_frame_validate (vm, r, next_index, n_vectors_left); nf = vlib_node_runtime_get_next_frame (vm, r, next_index); - f = vlib_get_frame (vm, nf->frame_index); + f = vlib_get_frame (vm, nf->frame); /* Make sure that magic number is still there. Otherwise, caller has overrun frame meta data. */ @@ -502,7 +502,7 @@ vlib_put_next_frame (vlib_main_t * vm, vec_add2 (nm->pending_frames, p, 1); - p->frame_index = nf->frame_index; + p->frame = nf->frame; p->node_runtime_index = nf->node_runtime_index; p->next_frame_index = nf - nm->next_frames; nf->flags |= VLIB_FRAME_PENDING; @@ -516,7 +516,7 @@ vlib_put_next_frame (vlib_main_t * vm, */ if (0 && r->thread_index != next_runtime->thread_index) { - nf->frame_index = ~0; + nf->frame = NULL; nf->flags &= ~(VLIB_FRAME_PENDING | VLIB_FRAME_IS_ALLOCATED); } } @@ -542,15 +542,17 @@ never_inline void vlib_node_runtime_sync_stats (vlib_main_t * vm, vlib_node_runtime_t * r, uword n_calls, uword n_vectors, uword n_clocks, - uword n_ticks) + uword n_ticks0, uword n_ticks1) { vlib_node_t *n = vlib_get_node (vm, r->node_index); n->stats_total.calls += n_calls + r->calls_since_last_overflow; n->stats_total.vectors += n_vectors + r->vectors_since_last_overflow; n->stats_total.clocks += n_clocks + r->clocks_since_last_overflow; - n->stats_total.perf_counter_ticks += n_ticks + - r->perf_counter_ticks_since_last_overflow; + n->stats_total.perf_counter0_ticks += n_ticks0 + + r->perf_counter0_ticks_since_last_overflow; + n->stats_total.perf_counter1_ticks += n_ticks1 + + r->perf_counter1_ticks_since_last_overflow; n->stats_total.perf_counter_vectors += n_vectors + r->perf_counter_vectors_since_last_overflow; n->stats_total.max_clock = r->max_clock; @@ -559,7 +561,8 @@ vlib_node_runtime_sync_stats (vlib_main_t * vm, r->calls_since_last_overflow = 0; r->vectors_since_last_overflow = 0; r->clocks_since_last_overflow = 0; - r->perf_counter_ticks_since_last_overflow = 0ULL; + r->perf_counter0_ticks_since_last_overflow = 0ULL; + r->perf_counter1_ticks_since_last_overflow = 0ULL; r->perf_counter_vectors_since_last_overflow = 0ULL; } @@ -567,12 +570,12 @@ always_inline void __attribute__ ((unused)) vlib_process_sync_stats (vlib_main_t * vm, vlib_process_t * p, uword n_calls, uword n_vectors, uword n_clocks, - uword n_ticks) + uword n_ticks0, uword n_ticks1) { vlib_node_runtime_t *rt = &p->node_runtime; vlib_node_t *n = vlib_get_node (vm, rt->node_index); vlib_node_runtime_sync_stats (vm, rt, n_calls, n_vectors, n_clocks, - n_ticks); + n_ticks0, n_ticks1); n->stats_total.suspends += p->n_suspends; p->n_suspends = 0; } @@ -598,7 +601,7 @@ vlib_node_sync_stats (vlib_main_t * vm, vlib_node_t * n) vec_elt_at_index (vm->node_main.nodes_by_type[n->type], n->runtime_index); - vlib_node_runtime_sync_stats (vm, rt, 0, 0, 0, 0); + vlib_node_runtime_sync_stats (vm, rt, 0, 0, 0, 0, 0); /* Sync up runtime next frame vector counters with main node structure. */ { @@ -619,27 +622,30 @@ vlib_node_runtime_update_stats (vlib_main_t * vm, vlib_node_runtime_t * node, uword n_calls, uword n_vectors, uword n_clocks, - uword n_ticks) + uword n_ticks0, uword n_ticks1) { u32 ca0, ca1, v0, v1, cl0, cl1, r; - u32 ptick0, ptick1, pvec0, pvec1; + u32 ptick00, ptick01, ptick10, ptick11, pvec0, pvec1; cl0 = cl1 = node->clocks_since_last_overflow; ca0 = ca1 = node->calls_since_last_overflow; v0 = v1 = node->vectors_since_last_overflow; - ptick0 = ptick1 = node->perf_counter_ticks_since_last_overflow; + ptick00 = ptick01 = node->perf_counter0_ticks_since_last_overflow; + ptick10 = ptick11 = node->perf_counter1_ticks_since_last_overflow; pvec0 = pvec1 = node->perf_counter_vectors_since_last_overflow; ca1 = ca0 + n_calls; v1 = v0 + n_vectors; cl1 = cl0 + n_clocks; - ptick1 = ptick0 + n_ticks; + ptick01 = ptick00 + n_ticks0; + ptick11 = ptick10 + n_ticks1; pvec1 = pvec0 + n_vectors; node->calls_since_last_overflow = ca1; node->clocks_since_last_overflow = cl1; node->vectors_since_last_overflow = v1; - node->perf_counter_ticks_since_last_overflow = ptick1; + node->perf_counter0_ticks_since_last_overflow = ptick01; + node->perf_counter1_ticks_since_last_overflow = ptick11; node->perf_counter_vectors_since_last_overflow = pvec1; node->max_clock_n = node->max_clock > n_clocks ? @@ -648,38 +654,42 @@ vlib_node_runtime_update_stats (vlib_main_t * vm, r = vlib_node_runtime_update_main_loop_vector_stats (vm, node, n_vectors); - if (PREDICT_FALSE (ca1 < ca0 || v1 < v0 || cl1 < cl0) || (ptick1 < ptick0) - || (pvec1 < pvec0)) + if (PREDICT_FALSE (ca1 < ca0 || v1 < v0 || cl1 < cl0) || (ptick01 < ptick00) + || (ptick11 < ptick10) || (pvec1 < pvec0)) { node->calls_since_last_overflow = ca0; node->clocks_since_last_overflow = cl0; node->vectors_since_last_overflow = v0; - node->perf_counter_ticks_since_last_overflow = ptick0; + node->perf_counter0_ticks_since_last_overflow = ptick00; + node->perf_counter1_ticks_since_last_overflow = ptick10; node->perf_counter_vectors_since_last_overflow = pvec0; vlib_node_runtime_sync_stats (vm, node, n_calls, n_vectors, n_clocks, - n_ticks); + n_ticks0, n_ticks1); } return r; } -static inline u64 -vlib_node_runtime_perf_counter (vlib_main_t * vm) +always_inline void +vlib_node_runtime_perf_counter (vlib_main_t * vm, u64 * pmc0, u64 * pmc1, + vlib_node_runtime_t * node, + vlib_frame_t * frame, int before_or_after) { - if (PREDICT_FALSE (vm->vlib_node_runtime_perf_counter_cb != 0)) - return ((*vm->vlib_node_runtime_perf_counter_cb) (vm)); - return 0ULL; + *pmc0 = 0; + *pmc1 = 0; + if (PREDICT_FALSE (vec_len (vm->vlib_node_runtime_perf_counter_cbs) != 0)) + clib_call_callbacks (vm->vlib_node_runtime_perf_counter_cbs, vm, pmc0, + pmc1, node, frame, before_or_after); } always_inline void vlib_process_update_stats (vlib_main_t * vm, vlib_process_t * p, - uword n_calls, uword n_vectors, uword n_clocks, - uword n_ticks) + uword n_calls, uword n_vectors, uword n_clocks) { vlib_node_runtime_update_stats (vm, &p->node_runtime, - n_calls, n_vectors, n_clocks, n_ticks); + n_calls, n_vectors, n_clocks, 0ULL, 0ULL); } static clib_error_t * @@ -904,18 +914,35 @@ vlib_elog_main_loop_event (vlib_main_t * vm, { vlib_main_t *evm = &vlib_global_main; elog_main_t *em = &evm->elog_main; + int enabled = evm->elog_trace_graph_dispatch | + evm->elog_trace_graph_circuit; - if (VLIB_ELOG_MAIN_LOOP && n_vectors) - elog_track (em, - /* event type */ - vec_elt_at_index (is_return - ? evm->node_return_elog_event_types - : evm->node_call_elog_event_types, - node_index), - /* track */ - (vm->thread_index ? &vlib_worker_threads[vm->thread_index]. - elog_track : &em->default_track), - /* data to log */ n_vectors); + if (PREDICT_FALSE (enabled && n_vectors)) + { + if (PREDICT_FALSE (!elog_is_enabled (em))) + { + evm->elog_trace_graph_dispatch = 0; + evm->elog_trace_graph_circuit = 0; + return; + } + if (PREDICT_TRUE + (evm->elog_trace_graph_dispatch || + (evm->elog_trace_graph_circuit && + node_index == evm->elog_trace_graph_circuit_node_index))) + { + elog_track (em, + /* event type */ + vec_elt_at_index (is_return + ? evm->node_return_elog_event_types + : evm->node_call_elog_event_types, + node_index), + /* track */ + (vm->thread_index ? + &vlib_worker_threads[vm->thread_index].elog_track + : &em->default_track), + /* data to log */ n_vectors); + } + } } #if VLIB_BUFFER_TRACE_TRAJECTORY > 0 @@ -944,6 +971,163 @@ add_trajectory_trace (vlib_buffer_t * b, u32 node_index) #endif } +u8 *format_vnet_buffer_flags (u8 * s, va_list * args) __attribute__ ((weak)); +u8 * +format_vnet_buffer_flags (u8 * s, va_list * args) +{ + s = format (s, "BUG STUB %s", __FUNCTION__); + return s; +} + +u8 *format_vnet_buffer_opaque (u8 * s, va_list * args) __attribute__ ((weak)); +u8 * +format_vnet_buffer_opaque (u8 * s, va_list * args) +{ + s = format (s, "BUG STUB %s", __FUNCTION__); + return s; +} + +u8 *format_vnet_buffer_opaque2 (u8 * s, va_list * args) + __attribute__ ((weak)); +u8 * +format_vnet_buffer_opaque2 (u8 * s, va_list * args) +{ + s = format (s, "BUG STUB %s", __FUNCTION__); + return s; +} + +static u8 * +format_buffer_metadata (u8 * s, va_list * args) +{ + vlib_buffer_t *b = va_arg (*args, vlib_buffer_t *); + + s = format (s, "flags: %U\n", format_vnet_buffer_flags, b); + s = format (s, "current_data: %d, current_length: %d\n", + (i32) (b->current_data), (i32) (b->current_length)); + s = format + (s, + "current_config_index/punt_reason: %d, flow_id: %x, next_buffer: %x\n", + b->current_config_index, b->flow_id, b->next_buffer); + s = + format (s, "error: %d, ref_count: %d, buffer_pool_index: %d\n", + (u32) (b->error), (u32) (b->ref_count), + (u32) (b->buffer_pool_index)); + s = + format (s, "trace_handle: 0x%x, len_not_first_buf: %d\n", b->trace_handle, + b->total_length_not_including_first_buffer); + return s; +} + +#define A(x) vec_add1(vm->pcap_buffer, (x)) + +static void +dispatch_pcap_trace (vlib_main_t * vm, + vlib_node_runtime_t * node, vlib_frame_t * frame) +{ + int i; + vlib_buffer_t *bufs[VLIB_FRAME_SIZE], **bufp, *b; + pcap_main_t *pm = &vlib_global_main.dispatch_pcap_main; + vlib_trace_main_t *tm = &vm->trace_main; + u32 capture_size; + vlib_node_t *n; + i32 n_left; + f64 time_now = vlib_time_now (vm); + u32 *from; + u8 *d; + u8 string_count; + + /* Input nodes don't have frames yet */ + if (frame == 0 || frame->n_vectors == 0) + return; + + from = vlib_frame_vector_args (frame); + vlib_get_buffers (vm, from, bufs, frame->n_vectors); + bufp = bufs; + + n = vlib_get_node (vm, node->node_index); + + for (i = 0; i < frame->n_vectors; i++) + { + if (PREDICT_TRUE (pm->n_packets_captured < pm->n_packets_to_capture)) + { + b = bufp[i]; + + vec_reset_length (vm->pcap_buffer); + string_count = 0; + + /* Version, flags */ + A ((u8) VLIB_PCAP_MAJOR_VERSION); + A ((u8) VLIB_PCAP_MINOR_VERSION); + A (0 /* string_count */ ); + A (n->protocol_hint); + + /* Buffer index (big endian) */ + A ((from[i] >> 24) & 0xff); + A ((from[i] >> 16) & 0xff); + A ((from[i] >> 8) & 0xff); + A ((from[i] >> 0) & 0xff); + + /* Node name, NULL-terminated ASCII */ + vm->pcap_buffer = format (vm->pcap_buffer, "%v%c", n->name, 0); + string_count++; + + vm->pcap_buffer = format (vm->pcap_buffer, "%U%c", + format_buffer_metadata, b, 0); + string_count++; + vm->pcap_buffer = format (vm->pcap_buffer, "%U%c", + format_vnet_buffer_opaque, b, 0); + string_count++; + vm->pcap_buffer = format (vm->pcap_buffer, "%U%c", + format_vnet_buffer_opaque2, b, 0); + string_count++; + + /* Is this packet traced? */ + if (PREDICT_FALSE (b->flags & VLIB_BUFFER_IS_TRACED)) + { + vlib_trace_header_t **h + = pool_elt_at_index (tm->trace_buffer_pool, + vlib_buffer_get_trace_index (b)); + + vm->pcap_buffer = format (vm->pcap_buffer, "%U%c", + format_vlib_trace, vm, h[0], 0); + string_count++; + } + + /* Save the string count */ + vm->pcap_buffer[2] = string_count; + + /* Figure out how many bytes in the pcap trace */ + capture_size = vec_len (vm->pcap_buffer) + + +vlib_buffer_length_in_chain (vm, b); + + clib_spinlock_lock_if_init (&pm->lock); + n_left = clib_min (capture_size, 16384); + d = pcap_add_packet (pm, time_now, n_left, capture_size); + + /* Copy the header */ + clib_memcpy_fast (d, vm->pcap_buffer, vec_len (vm->pcap_buffer)); + d += vec_len (vm->pcap_buffer); + + n_left = clib_min + (vlib_buffer_length_in_chain (vm, b), + (16384 - vec_len (vm->pcap_buffer))); + /* Copy the packet data */ + while (1) + { + u32 copy_length = clib_min ((u32) n_left, b->current_length); + clib_memcpy_fast (d, b->data + b->current_data, copy_length); + n_left -= b->current_length; + if (n_left <= 0) + break; + d += b->current_length; + ASSERT (b->flags & VLIB_BUFFER_NEXT_PRESENT); + b = vlib_get_buffer (vm, b->next_buffer); + } + clib_spinlock_unlock_if_init (&pm->lock); + } + } +} + static_always_inline u64 dispatch_node (vlib_main_t * vm, vlib_node_runtime_t * node, @@ -955,6 +1139,7 @@ dispatch_node (vlib_main_t * vm, u64 t; vlib_node_main_t *nm = &vm->node_main; vlib_next_frame_t *nf; + u64 pmc_before[2], pmc_after[2], pmc_delta[2]; if (CLIB_DEBUG > 0) { @@ -990,136 +1175,144 @@ dispatch_node (vlib_main_t * vm, vm->cpu_time_last_node_dispatch = last_time_stamp; - if (1 /* || vm->thread_index == node->thread_index */ ) - { - u64 pmc_before, pmc_delta; + vlib_elog_main_loop_event (vm, node->node_index, + last_time_stamp, frame ? frame->n_vectors : 0, + /* is_after */ 0); - vlib_elog_main_loop_event (vm, node->node_index, - last_time_stamp, - frame ? frame->n_vectors : 0, - /* is_after */ 0); - - /* - * To validate accounting: pmc_before = last_time_stamp - * perf ticks should equal clocks/pkt... - */ - pmc_before = vlib_node_runtime_perf_counter (vm); + vlib_node_runtime_perf_counter (vm, &pmc_before[0], &pmc_before[1], + node, frame, 0 /* before */ ); - /* - * Turn this on if you run into - * "bad monkey" contexts, and you want to know exactly - * which nodes they've visited... See ixge.c... - */ - if (VLIB_BUFFER_TRACE_TRAJECTORY && frame) + /* + * Turn this on if you run into + * "bad monkey" contexts, and you want to know exactly + * which nodes they've visited... See ixge.c... + */ + if (VLIB_BUFFER_TRACE_TRAJECTORY && frame) + { + int i; + u32 *from; + from = vlib_frame_vector_args (frame); + for (i = 0; i < frame->n_vectors; i++) { - int i; - u32 *from; - from = vlib_frame_vector_args (frame); - for (i = 0; i < frame->n_vectors; i++) - { - vlib_buffer_t *b = vlib_get_buffer (vm, from[i]); - add_trajectory_trace (b, node->node_index); - } - n = node->function (vm, node, frame); + vlib_buffer_t *b = vlib_get_buffer (vm, from[i]); + add_trajectory_trace (b, node->node_index); } - else - n = node->function (vm, node, frame); + if (PREDICT_FALSE (vm->dispatch_pcap_enable)) + dispatch_pcap_trace (vm, node, frame); + n = node->function (vm, node, frame); + } + else + { + if (PREDICT_FALSE (vm->dispatch_pcap_enable)) + dispatch_pcap_trace (vm, node, frame); + n = node->function (vm, node, frame); + } - t = clib_cpu_time_now (); + t = clib_cpu_time_now (); - /* - * To validate accounting: pmc_delta = t - pmc_before; - * perf ticks should equal clocks/pkt... - */ - pmc_delta = vlib_node_runtime_perf_counter (vm) - pmc_before; - - vlib_elog_main_loop_event (vm, node->node_index, t, n, /* is_after */ - 1); - - vm->main_loop_vectors_processed += n; - vm->main_loop_nodes_processed += n > 0; - - v = vlib_node_runtime_update_stats (vm, node, - /* n_calls */ 1, - /* n_vectors */ n, - /* n_clocks */ t - last_time_stamp, - pmc_delta /* PMC ticks */ ); - - /* When in interrupt mode and vector rate crosses threshold switch to - polling mode. */ - if ((dispatch_state == VLIB_NODE_STATE_INTERRUPT) - || (dispatch_state == VLIB_NODE_STATE_POLLING - && (node->flags - & VLIB_NODE_FLAG_SWITCH_FROM_INTERRUPT_TO_POLLING_MODE))) - { -#ifdef DISPATCH_NODE_ELOG_REQUIRED - ELOG_TYPE_DECLARE (e) = - { - .function = (char *) __FUNCTION__,.format = - "%s vector length %d, switching to %s",.format_args = - "T4i4t4",.n_enum_strings = 2,.enum_strings = - { - "interrupt", "polling",},}; - struct - { - u32 node_name, vector_length, is_polling; - } *ed; - vlib_worker_thread_t *w = vlib_worker_threads + vm->thread_index; -#endif + /* + * To validate accounting: pmc_delta = t - pmc_before; + * perf ticks should equal clocks/pkt... + */ + vlib_node_runtime_perf_counter (vm, &pmc_after[0], &pmc_after[1], node, + frame, 1 /* after */ ); + + pmc_delta[0] = pmc_after[0] - pmc_before[0]; + pmc_delta[1] = pmc_after[1] - pmc_before[1]; + + vlib_elog_main_loop_event (vm, node->node_index, t, n, 1 /* is_after */ ); + + vm->main_loop_vectors_processed += n; + vm->main_loop_nodes_processed += n > 0; + + v = vlib_node_runtime_update_stats (vm, node, + /* n_calls */ 1, + /* n_vectors */ n, + /* n_clocks */ t - last_time_stamp, + pmc_delta[0] /* PMC0 */ , + pmc_delta[1] /* PMC1 */ ); + + /* When in interrupt mode and vector rate crosses threshold switch to + polling mode. */ + if (PREDICT_FALSE ((dispatch_state == VLIB_NODE_STATE_INTERRUPT) + || (dispatch_state == VLIB_NODE_STATE_POLLING + && (node->flags + & + VLIB_NODE_FLAG_SWITCH_FROM_INTERRUPT_TO_POLLING_MODE)))) + { + /* *INDENT-OFF* */ + ELOG_TYPE_DECLARE (e) = + { + .function = (char *) __FUNCTION__, + .format = "%s vector length %d, switching to %s", + .format_args = "T4i4t4", + .n_enum_strings = 2, + .enum_strings = { + "interrupt", "polling", + }, + }; + /* *INDENT-ON* */ + struct + { + u32 node_name, vector_length, is_polling; + } *ed; - if ((dispatch_state == VLIB_NODE_STATE_INTERRUPT - && v >= nm->polling_threshold_vector_length) && - !(node->flags & - VLIB_NODE_FLAG_SWITCH_FROM_INTERRUPT_TO_POLLING_MODE)) + if ((dispatch_state == VLIB_NODE_STATE_INTERRUPT + && v >= nm->polling_threshold_vector_length) && + !(node->flags & + VLIB_NODE_FLAG_SWITCH_FROM_INTERRUPT_TO_POLLING_MODE)) + { + vlib_node_t *n = vlib_get_node (vm, node->node_index); + n->state = VLIB_NODE_STATE_POLLING; + node->state = VLIB_NODE_STATE_POLLING; + node->flags &= + ~VLIB_NODE_FLAG_SWITCH_FROM_POLLING_TO_INTERRUPT_MODE; + node->flags |= VLIB_NODE_FLAG_SWITCH_FROM_INTERRUPT_TO_POLLING_MODE; + nm->input_node_counts_by_state[VLIB_NODE_STATE_INTERRUPT] -= 1; + nm->input_node_counts_by_state[VLIB_NODE_STATE_POLLING] += 1; + + if (PREDICT_FALSE (vlib_global_main.elog_trace_graph_dispatch)) { - vlib_node_t *n = vlib_get_node (vm, node->node_index); - n->state = VLIB_NODE_STATE_POLLING; - node->state = VLIB_NODE_STATE_POLLING; - node->flags &= - ~VLIB_NODE_FLAG_SWITCH_FROM_POLLING_TO_INTERRUPT_MODE; - node->flags |= - VLIB_NODE_FLAG_SWITCH_FROM_INTERRUPT_TO_POLLING_MODE; - nm->input_node_counts_by_state[VLIB_NODE_STATE_INTERRUPT] -= 1; - nm->input_node_counts_by_state[VLIB_NODE_STATE_POLLING] += 1; + vlib_worker_thread_t *w = vlib_worker_threads + + vm->thread_index; -#ifdef DISPATCH_NODE_ELOG_REQUIRED ed = ELOG_TRACK_DATA (&vlib_global_main.elog_main, e, w->elog_track); ed->node_name = n->name_elog_string; ed->vector_length = v; ed->is_polling = 1; -#endif } - else if (dispatch_state == VLIB_NODE_STATE_POLLING - && v <= nm->interrupt_threshold_vector_length) + } + else if (dispatch_state == VLIB_NODE_STATE_POLLING + && v <= nm->interrupt_threshold_vector_length) + { + vlib_node_t *n = vlib_get_node (vm, node->node_index); + if (node->flags & + VLIB_NODE_FLAG_SWITCH_FROM_POLLING_TO_INTERRUPT_MODE) { - vlib_node_t *n = vlib_get_node (vm, node->node_index); - if (node->flags & - VLIB_NODE_FLAG_SWITCH_FROM_POLLING_TO_INTERRUPT_MODE) - { - /* Switch to interrupt mode after dispatch in polling one more time. - This allows driver to re-enable interrupts. */ - n->state = VLIB_NODE_STATE_INTERRUPT; - node->state = VLIB_NODE_STATE_INTERRUPT; - node->flags &= - ~VLIB_NODE_FLAG_SWITCH_FROM_INTERRUPT_TO_POLLING_MODE; - nm->input_node_counts_by_state[VLIB_NODE_STATE_POLLING] -= - 1; - nm->input_node_counts_by_state[VLIB_NODE_STATE_INTERRUPT] += - 1; + /* Switch to interrupt mode after dispatch in polling one more time. + This allows driver to re-enable interrupts. */ + n->state = VLIB_NODE_STATE_INTERRUPT; + node->state = VLIB_NODE_STATE_INTERRUPT; + node->flags &= + ~VLIB_NODE_FLAG_SWITCH_FROM_INTERRUPT_TO_POLLING_MODE; + nm->input_node_counts_by_state[VLIB_NODE_STATE_POLLING] -= 1; + nm->input_node_counts_by_state[VLIB_NODE_STATE_INTERRUPT] += 1; - } - else + } + else + { + vlib_worker_thread_t *w = vlib_worker_threads + + vm->thread_index; + node->flags |= + VLIB_NODE_FLAG_SWITCH_FROM_POLLING_TO_INTERRUPT_MODE; + if (PREDICT_FALSE (vlib_global_main.elog_trace_graph_dispatch)) { - node->flags |= - VLIB_NODE_FLAG_SWITCH_FROM_POLLING_TO_INTERRUPT_MODE; -#ifdef DISPATCH_NODE_ELOG_REQUIRED ed = ELOG_TRACK_DATA (&vlib_global_main.elog_main, e, w->elog_track); ed->node_name = n->name_elog_string; ed->vector_length = v; ed->is_polling = 0; -#endif } } } @@ -1136,7 +1329,7 @@ dispatch_pending_node (vlib_main_t * vm, uword pending_frame_index, vlib_frame_t *f; vlib_next_frame_t *nf, nf_dummy; vlib_node_runtime_t *n; - u32 restore_frame_index; + vlib_frame_t *restore_frame; vlib_pending_frame_t *p; /* See comment below about dangling references to nm->pending_frames */ @@ -1145,13 +1338,13 @@ dispatch_pending_node (vlib_main_t * vm, uword pending_frame_index, n = vec_elt_at_index (nm->nodes_by_type[VLIB_NODE_TYPE_INTERNAL], p->node_runtime_index); - f = vlib_get_frame (vm, p->frame_index); + f = vlib_get_frame (vm, p->frame); if (p->next_frame_index == VLIB_PENDING_FRAME_NO_NEXT_FRAME) { /* No next frame: so use dummy on stack. */ nf = &nf_dummy; nf->flags = f->frame_flags & VLIB_NODE_FLAG_TRACE; - nf->frame_index = ~p->frame_index; + nf->frame = NULL; } else nf = vec_elt_at_index (nm->next_frames, p->next_frame_index); @@ -1160,13 +1353,13 @@ dispatch_pending_node (vlib_main_t * vm, uword pending_frame_index, /* Force allocation of new frame while current frame is being dispatched. */ - restore_frame_index = ~0; - if (nf->frame_index == p->frame_index) + restore_frame = NULL; + if (nf->frame == p->frame) { - nf->frame_index = ~0; + nf->frame = NULL; nf->flags &= ~VLIB_FRAME_IS_ALLOCATED; if (!(n->flags & VLIB_NODE_FLAG_FRAME_NO_FREE_AFTER_DISPATCH)) - restore_frame_index = p->frame_index; + restore_frame = p->frame; } /* Frame must be pending. */ @@ -1184,11 +1377,17 @@ dispatch_pending_node (vlib_main_t * vm, uword pending_frame_index, VLIB_NODE_TYPE_INTERNAL, VLIB_NODE_STATE_POLLING, f, last_time_stamp); + /* Internal node vector-rate accounting, for summary stats */ + vm->internal_node_vectors += f->n_vectors; + vm->internal_node_calls++; + vm->internal_node_last_vectors_per_main_loop = + (f->n_vectors > vm->internal_node_last_vectors_per_main_loop) ? + f->n_vectors : vm->internal_node_last_vectors_per_main_loop; - f->frame_flags &= ~VLIB_FRAME_PENDING; + f->frame_flags &= ~(VLIB_FRAME_PENDING | VLIB_FRAME_NO_APPEND); /* Frame is ready to be used again, so restore it. */ - if (restore_frame_index != ~0) + if (restore_frame != NULL) { /* * We musn't restore a frame that is flagged to be freed. This @@ -1216,10 +1415,10 @@ dispatch_pending_node (vlib_main_t * vm, uword pending_frame_index, nf = vec_elt_at_index (nm->next_frames, p->next_frame_index); nf->flags |= VLIB_FRAME_IS_ALLOCATED; - if (~0 == nf->frame_index) + if (NULL == nf->frame) { /* no new frame has been assigned to this node, use the saved one */ - nf->frame_index = restore_frame_index; + nf->frame = restore_frame; f->n_vectors = 0; } else @@ -1353,7 +1552,7 @@ dispatch_process (vlib_main_t * vm, n_vectors = 0; pool_get (nm->suspended_process_frames, pf); pf->node_runtime_index = node->runtime_index; - pf->frame_index = f ? vlib_frame_index (vm, f) : ~0; + pf->frame = f; pf->next_frame_index = ~0; p->n_suspends += 1; @@ -1382,8 +1581,7 @@ dispatch_process (vlib_main_t * vm, vlib_process_update_stats (vm, p, /* n_calls */ !is_suspend, /* n_vectors */ n_vectors, - /* n_clocks */ t - last_time_stamp, - /* pmc_ticks */ 0ULL); + /* n_clocks */ t - last_time_stamp); return t; } @@ -1422,7 +1620,7 @@ dispatch_suspended_process (vlib_main_t * vm, node_runtime = &p->node_runtime; node = vlib_get_node (vm, node_runtime->node_index); - f = pf->frame_index != ~0 ? vlib_get_frame (vm, pf->frame_index) : 0; + f = pf->frame; vlib_elog_main_loop_event (vm, node_runtime->node_index, t, f ? f->n_vectors : 0, /* is_after */ 0); @@ -1466,8 +1664,7 @@ dispatch_suspended_process (vlib_main_t * vm, vlib_process_update_stats (vm, p, /* n_calls */ !is_suspend, /* n_vectors */ n_vectors, - /* n_clocks */ t - last_time_stamp, - /* pmc_ticks */ 0ULL); + /* n_clocks */ t - last_time_stamp); return t; } @@ -1488,6 +1685,7 @@ vlib_main_or_worker_loop (vlib_main_t * vm, int is_main) u64 cpu_time_now; vlib_frame_queue_main_t *fqm; u32 *last_node_runtime_indices = 0; + u32 frame_queue_check_counter = 0; /* Initialize pending node vector. */ if (is_main) @@ -1517,13 +1715,21 @@ vlib_main_or_worker_loop (vlib_main_t * vm, int is_main) if (!nm->interrupt_threshold_vector_length) nm->interrupt_threshold_vector_length = 5; - /* Make sure the performance monitor counter is disabled */ - vm->perf_counter_id = ~0; + vm->cpu_id = clib_get_current_cpu_id (); + vm->numa_node = clib_get_current_numa_node (); /* Start all processes. */ if (is_main) { uword i; + + /* + * Perform an initial barrier sync. Pays no attention to + * the barrier sync hold-down timer scheme, which won't work + * at this point in time. + */ + vlib_worker_thread_initial_barrier_sync_and_release (vm); + nm->current_process_index = ~0; for (i = 0; i < vec_len (nm->processes); i++) cpu_time_now = dispatch_process (vm, nm->processes[i], /* frame */ 0, @@ -1543,11 +1749,28 @@ vlib_main_or_worker_loop (vlib_main_t * vm, int is_main) if (!is_main) { vlib_worker_thread_barrier_check (); - vec_foreach (fqm, tm->frame_queue_mains) - vlib_frame_queue_dequeue (vm, fqm); - if (PREDICT_FALSE (vm->worker_thread_main_loop_callback != 0)) - ((void (*)(vlib_main_t *)) vm->worker_thread_main_loop_callback) - (vm); + if (PREDICT_FALSE (vm->check_frame_queues + + frame_queue_check_counter)) + { + u32 processed = 0; + + if (vm->check_frame_queues) + { + frame_queue_check_counter = 100; + vm->check_frame_queues = 0; + } + + vec_foreach (fqm, tm->frame_queue_mains) + processed += vlib_frame_queue_dequeue (vm, fqm); + + /* No handoff queue work found? */ + if (processed) + frame_queue_check_counter = 100; + else + frame_queue_check_counter--; + } + if (PREDICT_FALSE (vec_len (vm->worker_thread_main_loop_callbacks))) + clib_call_callbacks (vm->worker_thread_main_loop_callbacks, vm); } /* Process pre-input nodes. */ @@ -1613,9 +1836,30 @@ vlib_main_or_worker_loop (vlib_main_t * vm, int is_main) if (is_main) { + /* *INDENT-OFF* */ + ELOG_TYPE_DECLARE (es) = + { + .format = "process tw start", + .format_args = "", + }; + ELOG_TYPE_DECLARE (ee) = + { + .format = "process tw end: %d", + .format_args = "i4", + }; + /* *INDENT-ON* */ + + struct + { + int nready_procs; + } *ed; + /* Check if process nodes have expired from timing wheel. */ ASSERT (nm->data_from_advancing_timing_wheel != 0); + if (PREDICT_FALSE (vm->elog_trace_graph_dispatch)) + ed = ELOG_DATA (&vlib_global_main.elog_main, es); + nm->data_from_advancing_timing_wheel = TW (tw_timer_expire_timers_vec) ((TWT (tw_timer_wheel) *) nm->timing_wheel, vlib_time_now (vm), @@ -1623,6 +1867,13 @@ vlib_main_or_worker_loop (vlib_main_t * vm, int is_main) ASSERT (nm->data_from_advancing_timing_wheel != 0); + if (PREDICT_FALSE (vm->elog_trace_graph_dispatch)) + { + ed = ELOG_DATA (&vlib_global_main.elog_main, ee); + ed->nready_procs = + _vec_len (nm->data_from_advancing_timing_wheel); + } + if (PREDICT_FALSE (_vec_len (nm->data_from_advancing_timing_wheel) > 0)) { @@ -1650,12 +1901,12 @@ vlib_main_or_worker_loop (vlib_main_t * vm, int is_main) te->n_data_elts, te->n_data_elt_bytes); if (te->n_data_bytes < sizeof (te->inline_event_data)) - clib_memcpy (data, te->inline_event_data, - te->n_data_bytes); + clib_memcpy_fast (data, te->inline_event_data, + te->n_data_bytes); else { - clib_memcpy (data, te->event_data_as_vector, - te->n_data_bytes); + clib_memcpy_fast (data, te->event_data_as_vector, + te->n_data_bytes); vec_free (te->event_data_as_vector); } pool_put (nm->signal_timed_event_data_pool, te); @@ -1671,7 +1922,6 @@ vlib_main_or_worker_loop (vlib_main_t * vm, int is_main) } } vlib_increment_main_loop_counter (vm); - /* Record time stamp in case there are no enabled nodes and above calls do not update time stamp. */ cpu_time_now = clib_cpu_time_now (); @@ -1739,6 +1989,29 @@ clib_error_t *name (vlib_main_t *vm) { return 0; } foreach_weak_reference_stub; #undef _ +void vl_api_set_elog_main (elog_main_t * m) __attribute__ ((weak)); +void +vl_api_set_elog_main (elog_main_t * m) +{ + clib_warning ("STUB"); +} + +int vl_api_set_elog_trace_api_messages (int enable) __attribute__ ((weak)); +int +vl_api_set_elog_trace_api_messages (int enable) +{ + clib_warning ("STUB"); + return 0; +} + +int vl_api_get_elog_trace_api_messages (void) __attribute__ ((weak)); +int +vl_api_get_elog_trace_api_messages (void) +{ + clib_warning ("STUB"); + return 0; +} + /* Main function. */ int vlib_main (vlib_main_t * volatile vm, unformat_input_t * input) @@ -1755,6 +2028,8 @@ vlib_main (vlib_main_t * volatile vm, unformat_input_t * input) vm->elog_main.event_ring_size = 128 << 10; elog_init (&vm->elog_main, vm->elog_main.event_ring_size); elog_enable_disable (&vm->elog_main, 1); + vl_api_set_elog_main (&vm->elog_main); + (void) vl_api_set_elog_trace_api_messages (1); /* Default name. */ if (!vm->name) @@ -1766,19 +2041,19 @@ vlib_main (vlib_main_t * volatile vm, unformat_input_t * input) goto done; } - if ((error = vlib_buffer_main_init (vm))) + if ((error = vlib_map_stat_segment_init (vm))) { clib_error_report (error); goto done; } - if ((error = vlib_thread_init (vm))) + if ((error = vlib_buffer_main_init (vm))) { clib_error_report (error); goto done; } - if ((error = vlib_map_stat_segment_init (vm))) + if ((error = vlib_thread_init (vm))) { clib_error_report (error); goto done; @@ -1828,10 +2103,6 @@ vlib_main (vlib_main_t * volatile vm, unformat_input_t * input) if ((error = vlib_call_all_init_functions (vm))) goto done; - /* Create default buffer free list. */ - vlib_buffer_create_free_list (vm, VLIB_BUFFER_DEFAULT_FREE_LIST_BYTES, - "default"); - nm->timing_wheel = clib_mem_alloc_aligned (sizeof (TWT (tw_timer_wheel)), CLIB_CACHE_LINE_BYTES); @@ -1849,6 +2120,20 @@ vlib_main (vlib_main_t * volatile vm, unformat_input_t * input) vec_validate (vm->processing_rpc_requests, 0); _vec_len (vm->processing_rpc_requests) = 0; + if ((error = vlib_call_all_config_functions (vm, input, 0 /* is_early */ ))) + goto done; + + /* Sort per-thread init functions before we start threads */ + vlib_sort_init_exit_functions (&vm->worker_init_function_registrations); + + /* Call all main loop enter functions. */ + { + clib_error_t *sub_error; + sub_error = vlib_call_all_main_loop_enter_functions (vm); + if (sub_error) + clib_error_report (sub_error); + } + switch (clib_setjmp (&vm->main_loop_exit, VLIB_MAIN_LOOP_EXIT_NONE)) { case VLIB_MAIN_LOOP_EXIT_NONE: @@ -1863,17 +2148,6 @@ vlib_main (vlib_main_t * volatile vm, unformat_input_t * input) goto done; } - if ((error = vlib_call_all_config_functions (vm, input, 0 /* is_early */ ))) - goto done; - - /* Call all main loop enter functions. */ - { - clib_error_t *sub_error; - sub_error = vlib_call_all_main_loop_enter_functions (vm); - if (sub_error) - clib_error_report (sub_error); - } - vlib_main_loop (vm); done: @@ -1891,6 +2165,270 @@ done: return 0; } +int +vlib_pcap_dispatch_trace_configure (vlib_pcap_dispatch_trace_args_t * a) +{ + vlib_main_t *vm = vlib_get_main (); + pcap_main_t *pm = &vm->dispatch_pcap_main; + vlib_trace_main_t *tm; + vlib_trace_node_t *tn; + + if (a->status) + { + if (vm->dispatch_pcap_enable) + { + int i; + vlib_cli_output + (vm, "pcap dispatch capture enabled: %d of %d pkts...", + pm->n_packets_captured, pm->n_packets_to_capture); + vlib_cli_output (vm, "capture to file %s", pm->file_name); + + for (i = 0; i < vec_len (vm->dispatch_buffer_trace_nodes); i++) + { + vlib_cli_output (vm, + "Buffer trace of %d pkts from %U enabled...", + a->buffer_traces_to_capture, + format_vlib_node_name, vm, + vm->dispatch_buffer_trace_nodes[i]); + } + } + else + vlib_cli_output (vm, "pcap dispatch capture disabled"); + return 0; + } + + /* Consistency checks */ + + /* Enable w/ capture already enabled not allowed */ + if (vm->dispatch_pcap_enable && a->enable) + return -7; /* VNET_API_ERROR_INVALID_VALUE */ + + /* Disable capture with capture already disabled, not interesting */ + if (vm->dispatch_pcap_enable == 0 && a->enable == 0) + return -81; /* VNET_API_ERROR_VALUE_EXIST */ + + /* Change number of packets to capture while capturing */ + if (vm->dispatch_pcap_enable && a->enable + && (pm->n_packets_to_capture != a->packets_to_capture)) + return -8; /* VNET_API_ERROR_INVALID_VALUE_2 */ + + /* Independent of enable/disable, to allow buffer trace multi nodes */ + if (a->buffer_trace_node_index != ~0) + { + /* *INDENT-OFF* */ + foreach_vlib_main (( + { + tm = &this_vlib_main->trace_main; + tm->verbose = 0; /* not sure this ever did anything... */ + vec_validate (tm->nodes, a->buffer_trace_node_index); + tn = tm->nodes + a->buffer_trace_node_index; + tn->limit += a->buffer_traces_to_capture; + tm->trace_enable = 1; + })); + /* *INDENT-ON* */ + vec_add1 (vm->dispatch_buffer_trace_nodes, a->buffer_trace_node_index); + } + + if (a->enable) + { + /* Clean up from previous run, if any */ + vec_free (pm->file_name); + vec_free (pm->pcap_data); + memset (pm, 0, sizeof (*pm)); + + vec_validate_aligned (vnet_trace_dummy, 2048, CLIB_CACHE_LINE_BYTES); + if (pm->lock == 0) + clib_spinlock_init (&(pm->lock)); + + if (a->filename == 0) + a->filename = format (0, "/tmp/dispatch.pcap%c", 0); + + pm->file_name = (char *) a->filename; + pm->n_packets_captured = 0; + pm->packet_type = PCAP_PACKET_TYPE_vpp; + pm->n_packets_to_capture = a->packets_to_capture; + /* *INDENT-OFF* */ + foreach_vlib_main (({this_vlib_main->dispatch_pcap_enable = 1;})); + /* *INDENT-ON* */ + } + else + { + /* *INDENT-OFF* */ + foreach_vlib_main (({this_vlib_main->dispatch_pcap_enable = 0;})); + /* *INDENT-ON* */ + vec_reset_length (vm->dispatch_buffer_trace_nodes); + if (pm->n_packets_captured) + { + clib_error_t *error; + pm->n_packets_to_capture = pm->n_packets_captured; + vlib_cli_output (vm, "Write %d packets to %s, and stop capture...", + pm->n_packets_captured, pm->file_name); + error = pcap_write (pm); + if (pm->flags & PCAP_MAIN_INIT_DONE) + pcap_close (pm); + /* Report I/O errors... */ + if (error) + { + clib_error_report (error); + return -11; /* VNET_API_ERROR_SYSCALL_ERROR_1 */ + } + return 0; + } + else + return -6; /* VNET_API_ERROR_NO_SUCH_ENTRY */ + } + + return 0; +} + +static clib_error_t * +dispatch_trace_command_fn (vlib_main_t * vm, + unformat_input_t * input, vlib_cli_command_t * cmd) +{ + unformat_input_t _line_input, *line_input = &_line_input; + vlib_pcap_dispatch_trace_args_t _a, *a = &_a; + u8 *filename = 0; + u32 max = 1000; + int rv; + int enable = 0; + int status = 0; + u32 node_index = ~0, buffer_traces_to_capture = 100; + + /* Get a line of input. */ + if (!unformat_user (input, unformat_line_input, line_input)) + return 0; + + while (unformat_check_input (line_input) != UNFORMAT_END_OF_INPUT) + { + if (unformat (line_input, "on %=", &enable, 1)) + ; + else if (unformat (line_input, "enable %=", &enable, 1)) + ; + else if (unformat (line_input, "off %=", &enable, 0)) + ; + else if (unformat (line_input, "disable %=", &enable, 0)) + ; + else if (unformat (line_input, "max %d", &max)) + ; + else if (unformat (line_input, "packets-to-capture %d", &max)) + ; + else if (unformat (line_input, "file %U", unformat_vlib_tmpfile, + &filename)) + ; + else if (unformat (line_input, "status %=", &status, 1)) + ; + else if (unformat (line_input, "buffer-trace %U %d", + unformat_vlib_node, vm, &node_index, + &buffer_traces_to_capture)) + ; + else + { + return clib_error_return (0, "unknown input `%U'", + format_unformat_error, line_input); + } + } + + unformat_free (line_input); + + /* no need for memset (a, 0, sizeof (*a)), set all fields here. */ + a->filename = filename; + a->enable = enable; + a->status = status; + a->packets_to_capture = max; + a->buffer_trace_node_index = node_index; + a->buffer_traces_to_capture = buffer_traces_to_capture; + + rv = vlib_pcap_dispatch_trace_configure (a); + + switch (rv) + { + case 0: + break; + + case -7: + return clib_error_return (0, "dispatch trace already enabled..."); + + case -81: + return clib_error_return (0, "dispatch trace already disabled..."); + + case -8: + return clib_error_return + (0, "can't change number of records to capture while tracing..."); + + case -11: + return clib_error_return (0, "I/O writing trace capture..."); + + case -6: + return clib_error_return (0, "No packets captured..."); + + default: + vlib_cli_output (vm, "WARNING: trace configure returned %d", rv); + break; + } + return 0; +} + +/*? + * This command is used to start or stop pcap dispatch trace capture, or show + * the capture status. + * + * This command has the following optional parameters: + * + * - on|off - Used to start or stop capture. + * + * - max - Depth of local buffer. Once 'nn' number + * of packets have been received, buffer is flushed to file. Once another + * 'nn' number of packets have been received, buffer is flushed + * to file, overwriting previous write. If not entered, value defaults + * to 100. Can only be updated if packet capture is off. + * + * - file - Used to specify the output filename. The file will + * be placed in the '/tmp' directory, so only the filename is + * supported. Directory should not be entered. If file already exists, file + * will be overwritten. If no filename is provided, '/tmp/vpe.pcap' + * will be used. Can only be updated if packet capture is off. + * + * - status - Displays the current status and configured attributes + * associated with a packet capture. If packet capture is in progress, + * 'status' also will return the number of packets currently in + * the local buffer. All additional attributes entered on command line + * with 'status' will be ignored and not applied. + * + * @cliexpar + * Example of how to display the status of capture when off: + * @cliexstart{pcap dispatch trace status} + * max is 100, for any interface to file /tmp/vpe.pcap + * pcap dispatch capture is off... + * @cliexend + * Example of how to start a dispatch trace capture: + * @cliexstart{pcap dispatch trace on max 35 file dispatchTrace.pcap} + * pcap dispatch capture on... + * @cliexend + * Example of how to start a dispatch trace capture with buffer tracing + * @cliexstart{pcap dispatch trace on max 10000 file dispatchTrace.pcap buffer-trace dpdk-input 1000} + * pcap dispatch capture on... + * @cliexend + * Example of how to display the status of a tx packet capture in progress: + * @cliexstart{pcap tx trace status} + * max is 35, dispatch trace to file /tmp/vppTest.pcap + * pcap tx capture is on: 20 of 35 pkts... + * @cliexend + * Example of how to stop a tx packet capture: + * @cliexstart{vppctl pcap dispatch trace off} + * captured 21 pkts... + * saved to /tmp/dispatchTrace.pcap... + * @cliexend +?*/ +/* *INDENT-OFF* */ +VLIB_CLI_COMMAND (pcap_dispatch_trace_command, static) = { + .path = "pcap dispatch trace", + .short_help = + "pcap dispatch trace [on|off] [max ] [file ] [status]\n" + " [buffer-trace ]", + .function = dispatch_trace_command_fn, +}; +/* *INDENT-ON* */ + /* * fd.io coding-style-patch-verification: ON *