X-Git-Url: https://gerrit.fd.io/r/gitweb?a=blobdiff_plain;f=src%2Fvlib%2Fmain.c;h=50eebbaa4a586850d66a8d07cf4ec8f6a5e79e3f;hb=9aa4ac55b0a205e333e567f87c3cc9379af2363c;hp=73548fbea72834851acae9fc5054e665cfaae82c;hpb=04a7f05e91e919f51eaecaee476435484076655b;p=vpp.git diff --git a/src/vlib/main.c b/src/vlib/main.c index 73548fbea72..50eebbaa4a5 100644 --- a/src/vlib/main.c +++ b/src/vlib/main.c @@ -41,77 +41,20 @@ #include #include #include +#include #include #include -#include -CJ_GLOBAL_LOG_PROTOTYPE; - -/* Actually allocate a few extra slots of vector data to support - speculative vector enqueues which overflow vector data in next frame. */ -#define VLIB_FRAME_SIZE_ALLOC (VLIB_FRAME_SIZE + 4) - -u32 wraps; - -always_inline u32 -vlib_frame_bytes (u32 n_scalar_bytes, u32 n_vector_bytes) -{ - u32 n_bytes; - - /* Make room for vlib_frame_t plus scalar arguments. */ - n_bytes = vlib_frame_vector_byte_offset (n_scalar_bytes); - - /* Make room for vector arguments. - Allocate a few extra slots of vector data to support - speculative vector enqueues which overflow vector data in next frame. */ -#define VLIB_FRAME_SIZE_EXTRA 4 - n_bytes += (VLIB_FRAME_SIZE + VLIB_FRAME_SIZE_EXTRA) * n_vector_bytes; - - /* Magic number is first 32bit number after vector data. - Used to make sure that vector data is never overrun. */ #define VLIB_FRAME_MAGIC (0xabadc0ed) - n_bytes += sizeof (u32); - - /* Pad to cache line. */ - n_bytes = round_pow2 (n_bytes, CLIB_CACHE_LINE_BYTES); - - return n_bytes; -} always_inline u32 * vlib_frame_find_magic (vlib_frame_t * f, vlib_node_t * node) { - void *p = f; - - p += vlib_frame_vector_byte_offset (node->scalar_size); - - p += (VLIB_FRAME_SIZE + VLIB_FRAME_SIZE_EXTRA) * node->vector_size; - - return p; -} - -static vlib_frame_size_t * -get_frame_size_info (vlib_node_main_t * nm, - u32 n_scalar_bytes, u32 n_vector_bytes) -{ - uword key = (n_scalar_bytes << 16) | n_vector_bytes; - uword *p, i; - - p = hash_get (nm->frame_size_hash, key); - if (p) - i = p[0]; - else - { - i = vec_len (nm->frame_sizes); - vec_validate (nm->frame_sizes, i); - hash_set (nm->frame_size_hash, key, i); - } - - return vec_elt_at_index (nm->frame_sizes, i); + return (void *) f + node->magic_offset; } -static u32 +static vlib_frame_t * vlib_frame_alloc_to_node (vlib_main_t * vm, u32 to_node_index, u32 frame_flags) { @@ -119,38 +62,35 @@ vlib_frame_alloc_to_node (vlib_main_t * vm, u32 to_node_index, vlib_frame_size_t *fs; vlib_node_t *to_node; vlib_frame_t *f; - u32 fi, l, n, scalar_size, vector_size; + u32 l, n; + + ASSERT (vm == vlib_get_main ()); to_node = vlib_get_node (vm, to_node_index); - scalar_size = to_node->scalar_size; - vector_size = to_node->vector_size; + vec_validate (nm->frame_sizes, to_node->frame_size_index); + fs = vec_elt_at_index (nm->frame_sizes, to_node->frame_size_index); + + if (fs->frame_size == 0) + fs->frame_size = to_node->frame_size; + else + ASSERT (fs->frame_size == to_node->frame_size); - fs = get_frame_size_info (nm, scalar_size, vector_size); - n = vlib_frame_bytes (scalar_size, vector_size); - if ((l = vec_len (fs->free_frame_indices)) > 0) + n = fs->frame_size; + if ((l = vec_len (fs->free_frames)) > 0) { /* Allocate from end of free list. */ - fi = fs->free_frame_indices[l - 1]; - f = vlib_get_frame_no_check (vm, fi); - _vec_len (fs->free_frame_indices) = l - 1; + f = fs->free_frames[l - 1]; + vec_set_len (fs->free_frames, l - 1); } else { - f = clib_mem_alloc_aligned_no_fail (n, VLIB_FRAME_ALIGN); - f->thread_index = vm->thread_index; - fi = vlib_frame_index_no_check (vm, f); + f = clib_mem_alloc_aligned_no_fail (n, CLIB_CACHE_LINE_BYTES); } /* Poison frame when debugging. */ if (CLIB_DEBUG > 0) - { - u32 save_thread_index = f->thread_index; - - memset (f, 0xfe, n); - - f->thread_index = save_thread_index; - } + clib_memset_u8 (f, 0xfe, n); /* Insert magic number. */ { @@ -160,19 +100,22 @@ vlib_frame_alloc_to_node (vlib_main_t * vm, u32 to_node_index, *magic = VLIB_FRAME_MAGIC; } - f->flags = VLIB_FRAME_IS_ALLOCATED | frame_flags; + f->frame_flags = VLIB_FRAME_IS_ALLOCATED | frame_flags; f->n_vectors = 0; - f->scalar_size = scalar_size; - f->vector_size = vector_size; + f->scalar_offset = to_node->scalar_offset; + f->vector_offset = to_node->vector_offset; + f->aux_offset = to_node->aux_offset; + f->flags = 0; + f->frame_size_index = to_node->frame_size_index; fs->n_alloc_frames += 1; - return fi; + return f; } /* Allocate a frame for from FROM_NODE to TO_NODE via TO_NEXT_INDEX. Returns frame index. */ -static u32 +static vlib_frame_t * vlib_frame_alloc (vlib_main_t * vm, vlib_node_runtime_t * from_node_runtime, u32 to_next_index) { @@ -188,10 +131,35 @@ vlib_frame_alloc (vlib_main_t * vm, vlib_node_runtime_t * from_node_runtime, vlib_frame_t * vlib_get_frame_to_node (vlib_main_t * vm, u32 to_node_index) { - u32 fi = vlib_frame_alloc_to_node (vm, to_node_index, - /* frame_flags */ - VLIB_FRAME_FREE_AFTER_DISPATCH); - return vlib_get_frame (vm, fi); + vlib_frame_t *f = vlib_frame_alloc_to_node (vm, to_node_index, + /* frame_flags */ + VLIB_FRAME_FREE_AFTER_DISPATCH); + return vlib_get_frame (vm, f); +} + +static inline void +vlib_validate_frame_indices (vlib_frame_t * f) +{ + if (CLIB_DEBUG > 0) + { + int i; + u32 *from = vlib_frame_vector_args (f); + + /* Check for bad buffer index values */ + for (i = 0; i < f->n_vectors; i++) + { + if (from[i] == 0) + { + clib_warning ("BUG: buffer index 0 at index %d", i); + ASSERT (0); + } + else if (from[i] == 0xfefefefe) + { + clib_warning ("BUG: frame poison pattern at index %d", i); + ASSERT (0); + } + } + } } void @@ -203,45 +171,45 @@ vlib_put_frame_to_node (vlib_main_t * vm, u32 to_node_index, vlib_frame_t * f) if (f->n_vectors == 0) return; + ASSERT (vm == vlib_get_main ()); + + vlib_validate_frame_indices (f); + to_node = vlib_get_node (vm, to_node_index); vec_add2 (vm->node_main.pending_frames, p, 1); - f->flags |= VLIB_FRAME_PENDING; - p->frame_index = vlib_frame_index (vm, f); + f->frame_flags |= VLIB_FRAME_PENDING; + p->frame = vlib_get_frame (vm, f); p->node_runtime_index = to_node->runtime_index; p->next_frame_index = VLIB_PENDING_FRAME_NO_NEXT_FRAME; } /* Free given frame. */ void -vlib_frame_free (vlib_main_t * vm, vlib_node_runtime_t * r, vlib_frame_t * f) +vlib_frame_free (vlib_main_t *vm, vlib_frame_t *f) { vlib_node_main_t *nm = &vm->node_main; - vlib_node_t *node; vlib_frame_size_t *fs; - u32 frame_index; - - ASSERT (f->flags & VLIB_FRAME_IS_ALLOCATED); - node = vlib_get_node (vm, r->node_index); - fs = get_frame_size_info (nm, node->scalar_size, node->vector_size); + ASSERT (vm == vlib_get_main ()); + ASSERT (f->frame_flags & VLIB_FRAME_IS_ALLOCATED); - frame_index = vlib_frame_index (vm, f); + fs = vec_elt_at_index (nm->frame_sizes, f->frame_size_index); - ASSERT (f->flags & VLIB_FRAME_IS_ALLOCATED); + ASSERT (f->frame_flags & VLIB_FRAME_IS_ALLOCATED); /* No next frames may point to freed frame. */ if (CLIB_DEBUG > 0) { vlib_next_frame_t *nf; - vec_foreach (nf, vm->node_main.next_frames) - ASSERT (nf->frame_index != frame_index); + vec_foreach (nf, vm->node_main.next_frames) ASSERT (nf->frame != f); } - f->flags &= ~VLIB_FRAME_IS_ALLOCATED; + f->frame_flags &= ~(VLIB_FRAME_IS_ALLOCATED | VLIB_FRAME_NO_APPEND); + f->flags = 0; - vec_add1 (fs->free_frame_indices, frame_index); + vec_add1 (fs->free_frames, f); ASSERT (fs->n_alloc_frames > 0); fs->n_alloc_frames -= 1; } @@ -250,19 +218,24 @@ static clib_error_t * show_frame_stats (vlib_main_t * vm, unformat_input_t * input, vlib_cli_command_t * cmd) { - vlib_node_main_t *nm = &vm->node_main; vlib_frame_size_t *fs; - vlib_cli_output (vm, "%=6s%=12s%=12s", "Size", "# Alloc", "# Free"); - vec_foreach (fs, nm->frame_sizes) - { - u32 n_alloc = fs->n_alloc_frames; - u32 n_free = vec_len (fs->free_frame_indices); + vlib_cli_output (vm, "%=8s%=6s%=12s%=12s", "Thread", "Size", "# Alloc", + "# Free"); + foreach_vlib_main () + { + vlib_node_main_t *nm = &this_vlib_main->node_main; + vec_foreach (fs, nm->frame_sizes) + { + u32 n_alloc = fs->n_alloc_frames; + u32 n_free = vec_len (fs->free_frames); - if (n_alloc + n_free > 0) - vlib_cli_output (vm, "%=6d%=12d%=12d", - fs - nm->frame_sizes, n_alloc, n_free); - } + if (n_alloc + n_free > 0) + vlib_cli_output (vm, "%=8d%=6d%=12d%=12d", + this_vlib_main->thread_index, fs->frame_size, + n_alloc, n_free); + } + } return 0; } @@ -321,11 +294,11 @@ vlib_next_frame_change_ownership (vlib_main_t * vm, if (next_frame->flags & VLIB_FRAME_PENDING) { vlib_pending_frame_t *p; - if (next_frame->frame_index != ~0) + if (next_frame->frame != NULL) { vec_foreach (p, nm->pending_frames) { - if (p->frame_index == next_frame->frame_index) + if (p->frame == next_frame->frame) { p->next_frame_index = next_frame - vm->node_main.next_frames; @@ -377,35 +350,38 @@ vlib_get_next_frame_internal (vlib_main_t * vm, /* ??? Don't need valid flag: can use frame_index == ~0 */ if (PREDICT_FALSE (!(nf->flags & VLIB_FRAME_IS_ALLOCATED))) { - nf->frame_index = vlib_frame_alloc (vm, node, next_index); + nf->frame = vlib_frame_alloc (vm, node, next_index); nf->flags |= VLIB_FRAME_IS_ALLOCATED; } - f = vlib_get_frame (vm, nf->frame_index); + f = nf->frame; /* Has frame been removed from pending vector (e.g. finished dispatching)? If so we can reuse frame. */ - if ((nf->flags & VLIB_FRAME_PENDING) && !(f->flags & VLIB_FRAME_PENDING)) + if ((nf->flags & VLIB_FRAME_PENDING) + && !(f->frame_flags & VLIB_FRAME_PENDING)) { nf->flags &= ~VLIB_FRAME_PENDING; f->n_vectors = 0; + f->flags = 0; } - /* Allocate new frame if current one is already full. */ + /* Allocate new frame if current one is marked as no-append or + it is already full. */ n_used = f->n_vectors; - if (n_used >= VLIB_FRAME_SIZE || (allocate_new_next_frame && n_used > 0)) + if (n_used >= VLIB_FRAME_SIZE || (allocate_new_next_frame && n_used > 0) || + (f->frame_flags & VLIB_FRAME_NO_APPEND)) { /* Old frame may need to be freed after dispatch, since we'll have two redundant frames from node -> next node. */ if (!(nf->flags & VLIB_FRAME_NO_FREE_AFTER_DISPATCH)) { - vlib_frame_t *f_old = vlib_get_frame (vm, nf->frame_index); - f_old->flags |= VLIB_FRAME_FREE_AFTER_DISPATCH; + vlib_frame_t *f_old = vlib_get_frame (vm, nf->frame); + f_old->frame_flags |= VLIB_FRAME_FREE_AFTER_DISPATCH; } /* Allocate new frame to replace full one. */ - nf->frame_index = vlib_frame_alloc (vm, node, next_index); - f = vlib_get_frame (vm, nf->frame_index); + f = nf->frame = vlib_frame_alloc (vm, node, next_index); n_used = f->n_vectors; } @@ -434,9 +410,12 @@ vlib_put_next_frame_validate (vlib_main_t * vm, u32 n_before, n_after; nf = vlib_node_runtime_get_next_frame (vm, rt, next_index); - f = vlib_get_frame (vm, nf->frame_index); + f = vlib_get_frame (vm, nf->frame); ASSERT (n_vectors_left <= VLIB_FRAME_SIZE); + + vlib_validate_frame_indices (f); + n_after = VLIB_FRAME_SIZE - n_vectors_left; n_before = f->n_vectors; @@ -467,11 +446,11 @@ vlib_put_next_frame (vlib_main_t * vm, vlib_frame_t *f; u32 n_vectors_in_frame; - if (vm->buffer_main->callbacks_registered == 0 && CLIB_DEBUG > 0) + if (CLIB_DEBUG > 0) vlib_put_next_frame_validate (vm, r, next_index, n_vectors_left); nf = vlib_node_runtime_get_next_frame (vm, r, next_index); - f = vlib_get_frame (vm, nf->frame_index); + f = vlib_get_frame (vm, nf->frame); /* Make sure that magic number is still there. Otherwise, caller has overrun frame meta data. */ @@ -495,35 +474,19 @@ vlib_put_next_frame (vlib_main_t * vm, r->cached_next_index = next_index; - if (!(f->flags & VLIB_FRAME_PENDING)) + if (!(f->frame_flags & VLIB_FRAME_PENDING)) { __attribute__ ((unused)) vlib_node_t *node; - vlib_node_t *next_node; - vlib_node_runtime_t *next_runtime; node = vlib_get_node (vm, r->node_index); - next_node = vlib_get_next_node (vm, r->node_index, next_index); - next_runtime = vlib_node_get_runtime (vm, next_node->index); vec_add2 (nm->pending_frames, p, 1); - p->frame_index = nf->frame_index; + p->frame = nf->frame; p->node_runtime_index = nf->node_runtime_index; p->next_frame_index = nf - nm->next_frames; nf->flags |= VLIB_FRAME_PENDING; - f->flags |= VLIB_FRAME_PENDING; - - /* - * If we're going to dispatch this frame on another thread, - * force allocation of a new frame. Otherwise, we create - * a dangling frame reference. Each thread has its own copy of - * the next_frames vector. - */ - if (0 && r->thread_index != next_runtime->thread_index) - { - nf->frame_index = ~0; - nf->flags &= ~(VLIB_FRAME_PENDING | VLIB_FRAME_IS_ALLOCATED); - } + f->frame_flags |= VLIB_FRAME_PENDING; } /* Copy trace flag from next_frame and from runtime. */ @@ -543,13 +506,11 @@ vlib_put_next_frame (vlib_main_t * vm, } /* Sync up runtime (32 bit counters) and main node stats (64 bit counters). */ -never_inline void -vlib_node_runtime_sync_stats (vlib_main_t * vm, - vlib_node_runtime_t * r, - uword n_calls, uword n_vectors, uword n_clocks) +void +vlib_node_runtime_sync_stats_node (vlib_node_t *n, vlib_node_runtime_t *r, + uword n_calls, uword n_vectors, + uword n_clocks) { - vlib_node_t *n = vlib_get_node (vm, r->node_index); - n->stats_total.calls += n_calls + r->calls_since_last_overflow; n->stats_total.vectors += n_vectors + r->vectors_since_last_overflow; n->stats_total.clocks += n_clocks + r->clocks_since_last_overflow; @@ -561,6 +522,14 @@ vlib_node_runtime_sync_stats (vlib_main_t * vm, r->clocks_since_last_overflow = 0; } +void +vlib_node_runtime_sync_stats (vlib_main_t *vm, vlib_node_runtime_t *r, + uword n_calls, uword n_vectors, uword n_clocks) +{ + vlib_node_t *n = vlib_get_node (vm, r->node_index); + vlib_node_runtime_sync_stats_node (n, r, n_calls, n_vectors, n_clocks); +} + always_inline void __attribute__ ((unused)) vlib_process_sync_stats (vlib_main_t * vm, vlib_process_t * p, @@ -581,7 +550,7 @@ vlib_node_sync_stats (vlib_main_t * vm, vlib_node_t * n) if (n->type == VLIB_NODE_TYPE_PROCESS) { /* Nothing to do for PROCESS nodes except in main thread */ - if (vm != &vlib_global_main) + if (vm != vlib_get_first_main ()) return; vlib_process_t *p = vlib_get_process_from_node (vm, n); @@ -629,6 +598,7 @@ vlib_node_runtime_update_stats (vlib_main_t * vm, node->calls_since_last_overflow = ca1; node->clocks_since_last_overflow = cl1; node->vectors_since_last_overflow = v1; + node->max_clock_n = node->max_clock > n_clocks ? node->max_clock_n : n_vectors; node->max_clock = node->max_clock > n_clocks ? node->max_clock : n_clocks; @@ -640,6 +610,7 @@ vlib_node_runtime_update_stats (vlib_main_t * vm, node->calls_since_last_overflow = ca0; node->clocks_since_last_overflow = cl0; node->vectors_since_last_overflow = v0; + vlib_node_runtime_sync_stats (vm, node, n_calls, n_vectors, n_clocks); } @@ -659,7 +630,7 @@ static clib_error_t * vlib_cli_elog_clear (vlib_main_t * vm, unformat_input_t * input, vlib_cli_command_t * cmd) { - elog_reset_buffer (&vm->elog_main); + elog_reset_buffer (&vlib_global_main.elog_main); return 0; } @@ -676,7 +647,7 @@ static clib_error_t * elog_save_buffer (vlib_main_t * vm, unformat_input_t * input, vlib_cli_command_t * cmd) { - elog_main_t *em = &vm->elog_main; + elog_main_t *em = &vlib_global_main.elog_main; char *file, *chroot_file; clib_error_t *error = 0; @@ -710,21 +681,12 @@ elog_save_buffer (vlib_main_t * vm, } void -elog_post_mortem_dump (void) +vlib_post_mortem_dump (void) { - vlib_main_t *vm = &vlib_global_main; - elog_main_t *em = &vm->elog_main; - u8 *filename; - clib_error_t *error; + vlib_global_main_t *vgm = vlib_get_global_main (); - if (!vm->elog_post_mortem_dump) - return; - - filename = format (0, "/tmp/elog_post_mortem.%d%c", getpid (), 0); - error = elog_write_file (em, (char *) filename, 1 /* flush ring */ ); - if (error) - clib_error_report (error); - vec_free (filename); + for (int i = 0; i < vec_len (vgm->post_mortem_callbacks); i++) + (vgm->post_mortem_callbacks[i]) (); } /* *INDENT-OFF* */ @@ -739,7 +701,7 @@ static clib_error_t * elog_stop (vlib_main_t * vm, unformat_input_t * input, vlib_cli_command_t * cmd) { - elog_main_t *em = &vm->elog_main; + elog_main_t *em = &vlib_global_main.elog_main; em->n_total_events_disable_limit = em->n_total_events; @@ -759,7 +721,7 @@ static clib_error_t * elog_restart (vlib_main_t * vm, unformat_input_t * input, vlib_cli_command_t * cmd) { - elog_main_t *em = &vm->elog_main; + elog_main_t *em = &vlib_global_main.elog_main; em->n_total_events_disable_limit = ~0; @@ -776,14 +738,14 @@ VLIB_CLI_COMMAND (elog_restart_cli, static) = { /* *INDENT-ON* */ static clib_error_t * -elog_resize (vlib_main_t * vm, - unformat_input_t * input, vlib_cli_command_t * cmd) +elog_resize_command_fn (vlib_main_t * vm, + unformat_input_t * input, vlib_cli_command_t * cmd) { - elog_main_t *em = &vm->elog_main; + elog_main_t *em = &vlib_global_main.elog_main; u32 tmp; /* Stop the parade */ - elog_reset_buffer (&vm->elog_main); + elog_reset_buffer (em); if (unformat (input, "%d", &tmp)) { @@ -801,7 +763,7 @@ elog_resize (vlib_main_t * vm, VLIB_CLI_COMMAND (elog_resize_cli, static) = { .path = "event-logger resize", .short_help = "event-logger resize ", - .function = elog_resize, + .function = elog_resize_command_fn, }; /* *INDENT-ON* */ @@ -810,7 +772,7 @@ VLIB_CLI_COMMAND (elog_resize_cli, static) = { static void elog_show_buffer_internal (vlib_main_t * vm, u32 n_events_to_show) { - elog_main_t *em = &vm->elog_main; + elog_main_t *em = &vlib_global_main.elog_main; elog_event_t *e, *es; f64 dt; @@ -875,67 +837,49 @@ vlib_elog_main_loop_event (vlib_main_t * vm, u32 node_index, u64 time, u32 n_vectors, u32 is_return) { - vlib_main_t *evm = &vlib_global_main; - elog_main_t *em = &evm->elog_main; - - if (VLIB_ELOG_MAIN_LOOP && n_vectors) - elog_track (em, - /* event type */ - vec_elt_at_index (is_return - ? evm->node_return_elog_event_types - : evm->node_call_elog_event_types, - node_index), - /* track */ - (vm->thread_index ? &vlib_worker_threads[vm->thread_index]. - elog_track : &em->default_track), - /* data to log */ n_vectors); -} - -void -vlib_dump_context_trace (vlib_main_t * vm, u32 bi) -{ - vlib_node_main_t *vnm = &vm->node_main; - vlib_buffer_t *b; - u8 i, n; + vlib_main_t *evm = vlib_get_first_main (); + elog_main_t *em = vlib_get_elog_main (); + int enabled = evm->elog_trace_graph_dispatch | + evm->elog_trace_graph_circuit; - if (VLIB_BUFFER_TRACE_TRAJECTORY) + if (PREDICT_FALSE (enabled && n_vectors)) { - b = vlib_get_buffer (vm, bi); - n = b->pre_data[0]; - - fformat (stderr, "Context trace for bi %d b 0x%llx, visited %d\n", - bi, b, n); - - if (n == 0 || n > 20) + if (PREDICT_FALSE (!elog_is_enabled (em))) { - fformat (stderr, "n is unreasonable\n"); + evm->elog_trace_graph_dispatch = 0; + evm->elog_trace_graph_circuit = 0; return; } - - - for (i = 0; i < n; i++) + if (PREDICT_TRUE + (evm->elog_trace_graph_dispatch || + (evm->elog_trace_graph_circuit && + node_index == evm->elog_trace_graph_circuit_node_index))) { - u32 node_index; - - node_index = b->pre_data[i + 1]; - - if (node_index > vec_len (vnm->nodes)) - { - fformat (stderr, "Skip bogus node index %d\n", node_index); - continue; - } - - fformat (stderr, "%v (%d)\n", vnm->nodes[node_index]->name, - node_index); + elog_track (em, + /* event type */ + vec_elt_at_index (is_return + ? evm->node_return_elog_event_types + : evm->node_call_elog_event_types, + node_index), + /* track */ + (vm->thread_index ? + &vlib_worker_threads[vm->thread_index].elog_track + : &em->default_track), + /* data to log */ n_vectors); } } - else - { - fformat (stderr, - "in vlib/buffers.h, #define VLIB_BUFFER_TRACE_TRAJECTORY 1\n"); - } } +static inline void +add_trajectory_trace (vlib_buffer_t * b, u32 node_index) +{ +#if VLIB_BUFFER_TRACE_TRAJECTORY > 0 + if (PREDICT_FALSE (b->trajectory_nb >= VLIB_BUFFER_TRACE_TRAJECTORY_MAX)) + return; + b->trajectory_trace[b->trajectory_nb] = node_index; + b->trajectory_nb++; +#endif +} static_always_inline u64 dispatch_node (vlib_main_t * vm, @@ -983,128 +927,135 @@ dispatch_node (vlib_main_t * vm, vm->cpu_time_last_node_dispatch = last_time_stamp; - if (1 /* || vm->thread_index == node->thread_index */ ) - { - vlib_main_t *stat_vm; - - stat_vm = /* vlib_mains ? vlib_mains[0] : */ vm; + vlib_elog_main_loop_event (vm, node->node_index, + last_time_stamp, frame ? frame->n_vectors : 0, + /* is_after */ 0); - vlib_elog_main_loop_event (vm, node->node_index, - last_time_stamp, - frame ? frame->n_vectors : 0, - /* is_after */ 0); + vlib_node_runtime_perf_counter (vm, node, frame, 0, last_time_stamp, + VLIB_NODE_RUNTIME_PERF_BEFORE); - /* - * Turn this on if you run into - * "bad monkey" contexts, and you want to know exactly - * which nodes they've visited... See ixge.c... - */ - if (VLIB_BUFFER_TRACE_TRAJECTORY && frame) + /* + * Turn this on if you run into + * "bad monkey" contexts, and you want to know exactly + * which nodes they've visited... See ixge.c... + */ + if (VLIB_BUFFER_TRACE_TRAJECTORY && frame) + { + int i; + u32 *from; + from = vlib_frame_vector_args (frame); + for (i = 0; i < frame->n_vectors; i++) { - int i; - int log_index; - u32 *from; - from = vlib_frame_vector_args (frame); - for (i = 0; i < frame->n_vectors; i++) - { - vlib_buffer_t *b = vlib_get_buffer (vm, from[i]); - ASSERT (b->pre_data[0] < 32); - log_index = b->pre_data[0]++ + 1; - b->pre_data[log_index] = node->node_index; - } - n = node->function (vm, node, frame); + vlib_buffer_t *b = vlib_get_buffer (vm, from[i]); + add_trajectory_trace (b, node->node_index); } + if (PREDICT_TRUE (vm->dispatch_wrapper_fn == 0)) + n = node->function (vm, node, frame); else + n = vm->dispatch_wrapper_fn (vm, node, frame); + } + else + { + if (PREDICT_TRUE (vm->dispatch_wrapper_fn == 0)) n = node->function (vm, node, frame); + else + n = vm->dispatch_wrapper_fn (vm, node, frame); + } - t = clib_cpu_time_now (); + t = clib_cpu_time_now (); - vlib_elog_main_loop_event (vm, node->node_index, t, n, /* is_after */ - 1); + vlib_node_runtime_perf_counter (vm, node, frame, n, t, + VLIB_NODE_RUNTIME_PERF_AFTER); - vm->main_loop_vectors_processed += n; - vm->main_loop_nodes_processed += n > 0; + vlib_elog_main_loop_event (vm, node->node_index, t, n, 1 /* is_after */ ); - v = vlib_node_runtime_update_stats (stat_vm, node, - /* n_calls */ 1, - /* n_vectors */ n, - /* n_clocks */ t - last_time_stamp); + vm->main_loop_vectors_processed += n; + vm->main_loop_nodes_processed += n > 0; - /* When in interrupt mode and vector rate crosses threshold switch to - polling mode. */ - if ((dispatch_state == VLIB_NODE_STATE_INTERRUPT) - || (dispatch_state == VLIB_NODE_STATE_POLLING - && (node->flags - & VLIB_NODE_FLAG_SWITCH_FROM_INTERRUPT_TO_POLLING_MODE))) - { -#ifdef DISPATCH_NODE_ELOG_REQUIRED - ELOG_TYPE_DECLARE (e) = - { - .function = (char *) __FUNCTION__,.format = - "%s vector length %d, switching to %s",.format_args = - "T4i4t4",.n_enum_strings = 2,.enum_strings = - { - "interrupt", "polling",},}; - struct - { - u32 node_name, vector_length, is_polling; - } *ed; - vlib_worker_thread_t *w = vlib_worker_threads + vm->thread_index; -#endif + v = vlib_node_runtime_update_stats (vm, node, + /* n_calls */ 1, + /* n_vectors */ n, + /* n_clocks */ t - last_time_stamp); + + /* When in adaptive mode and vector rate crosses threshold switch to + polling mode and vice versa. */ + if (PREDICT_FALSE (node->flags & VLIB_NODE_FLAG_ADAPTIVE_MODE)) + { + /* *INDENT-OFF* */ + ELOG_TYPE_DECLARE (e) = + { + .function = (char *) __FUNCTION__, + .format = "%s vector length %d, switching to %s", + .format_args = "T4i4t4", + .n_enum_strings = 2, + .enum_strings = { + "interrupt", "polling", + }, + }; + /* *INDENT-ON* */ + struct + { + u32 node_name, vector_length, is_polling; + } *ed; - if ((dispatch_state == VLIB_NODE_STATE_INTERRUPT - && v >= nm->polling_threshold_vector_length) && - !(node->flags & - VLIB_NODE_FLAG_SWITCH_FROM_INTERRUPT_TO_POLLING_MODE)) + if ((dispatch_state == VLIB_NODE_STATE_INTERRUPT + && v >= nm->polling_threshold_vector_length) && + !(node->flags & + VLIB_NODE_FLAG_SWITCH_FROM_INTERRUPT_TO_POLLING_MODE)) + { + vlib_node_t *n = vlib_get_node (vm, node->node_index); + n->state = VLIB_NODE_STATE_POLLING; + node->state = VLIB_NODE_STATE_POLLING; + node->flags &= + ~VLIB_NODE_FLAG_SWITCH_FROM_POLLING_TO_INTERRUPT_MODE; + node->flags |= VLIB_NODE_FLAG_SWITCH_FROM_INTERRUPT_TO_POLLING_MODE; + nm->input_node_counts_by_state[VLIB_NODE_STATE_INTERRUPT] -= 1; + nm->input_node_counts_by_state[VLIB_NODE_STATE_POLLING] += 1; + + if (PREDICT_FALSE ( + vlib_get_first_main ()->elog_trace_graph_dispatch)) { - vlib_node_t *n = vlib_get_node (vm, node->node_index); - n->state = VLIB_NODE_STATE_POLLING; - node->state = VLIB_NODE_STATE_POLLING; - node->flags &= - ~VLIB_NODE_FLAG_SWITCH_FROM_POLLING_TO_INTERRUPT_MODE; - node->flags |= - VLIB_NODE_FLAG_SWITCH_FROM_INTERRUPT_TO_POLLING_MODE; - nm->input_node_counts_by_state[VLIB_NODE_STATE_INTERRUPT] -= 1; - nm->input_node_counts_by_state[VLIB_NODE_STATE_POLLING] += 1; + vlib_worker_thread_t *w = vlib_worker_threads + + vm->thread_index; -#ifdef DISPATCH_NODE_ELOG_REQUIRED ed = ELOG_TRACK_DATA (&vlib_global_main.elog_main, e, w->elog_track); ed->node_name = n->name_elog_string; ed->vector_length = v; ed->is_polling = 1; -#endif } - else if (dispatch_state == VLIB_NODE_STATE_POLLING - && v <= nm->interrupt_threshold_vector_length) + } + else if (dispatch_state == VLIB_NODE_STATE_POLLING + && v <= nm->interrupt_threshold_vector_length) + { + vlib_node_t *n = vlib_get_node (vm, node->node_index); + if (node->flags & + VLIB_NODE_FLAG_SWITCH_FROM_POLLING_TO_INTERRUPT_MODE) { - vlib_node_t *n = vlib_get_node (vm, node->node_index); - if (node->flags & - VLIB_NODE_FLAG_SWITCH_FROM_POLLING_TO_INTERRUPT_MODE) - { - /* Switch to interrupt mode after dispatch in polling one more time. - This allows driver to re-enable interrupts. */ - n->state = VLIB_NODE_STATE_INTERRUPT; - node->state = VLIB_NODE_STATE_INTERRUPT; - node->flags &= - ~VLIB_NODE_FLAG_SWITCH_FROM_INTERRUPT_TO_POLLING_MODE; - nm->input_node_counts_by_state[VLIB_NODE_STATE_POLLING] -= - 1; - nm->input_node_counts_by_state[VLIB_NODE_STATE_INTERRUPT] += - 1; + /* Switch to interrupt mode after dispatch in polling one more time. + This allows driver to re-enable interrupts. */ + n->state = VLIB_NODE_STATE_INTERRUPT; + node->state = VLIB_NODE_STATE_INTERRUPT; + node->flags &= + ~VLIB_NODE_FLAG_SWITCH_FROM_INTERRUPT_TO_POLLING_MODE; + nm->input_node_counts_by_state[VLIB_NODE_STATE_POLLING] -= 1; + nm->input_node_counts_by_state[VLIB_NODE_STATE_INTERRUPT] += 1; - } - else + } + else + { + vlib_worker_thread_t *w = vlib_worker_threads + + vm->thread_index; + node->flags |= + VLIB_NODE_FLAG_SWITCH_FROM_POLLING_TO_INTERRUPT_MODE; + if (PREDICT_FALSE ( + vlib_get_first_main ()->elog_trace_graph_dispatch)) { - node->flags |= - VLIB_NODE_FLAG_SWITCH_FROM_POLLING_TO_INTERRUPT_MODE; -#ifdef DISPATCH_NODE_ELOG_REQUIRED ed = ELOG_TRACK_DATA (&vlib_global_main.elog_main, e, w->elog_track); ed->node_name = n->name_elog_string; ed->vector_length = v; ed->is_polling = 0; -#endif } } } @@ -1119,9 +1070,9 @@ dispatch_pending_node (vlib_main_t * vm, uword pending_frame_index, { vlib_node_main_t *nm = &vm->node_main; vlib_frame_t *f; - vlib_next_frame_t *nf, nf_dummy; + vlib_next_frame_t *nf, nf_placeholder; vlib_node_runtime_t *n; - u32 restore_frame_index; + vlib_frame_t *restore_frame; vlib_pending_frame_t *p; /* See comment below about dangling references to nm->pending_frames */ @@ -1130,32 +1081,32 @@ dispatch_pending_node (vlib_main_t * vm, uword pending_frame_index, n = vec_elt_at_index (nm->nodes_by_type[VLIB_NODE_TYPE_INTERNAL], p->node_runtime_index); - f = vlib_get_frame (vm, p->frame_index); + f = vlib_get_frame (vm, p->frame); if (p->next_frame_index == VLIB_PENDING_FRAME_NO_NEXT_FRAME) { - /* No next frame: so use dummy on stack. */ - nf = &nf_dummy; - nf->flags = f->flags & VLIB_NODE_FLAG_TRACE; - nf->frame_index = ~p->frame_index; + /* No next frame: so use placeholder on stack. */ + nf = &nf_placeholder; + nf->flags = f->frame_flags & VLIB_NODE_FLAG_TRACE; + nf->frame = NULL; } else nf = vec_elt_at_index (nm->next_frames, p->next_frame_index); - ASSERT (f->flags & VLIB_FRAME_IS_ALLOCATED); + ASSERT (f->frame_flags & VLIB_FRAME_IS_ALLOCATED); /* Force allocation of new frame while current frame is being dispatched. */ - restore_frame_index = ~0; - if (nf->frame_index == p->frame_index) + restore_frame = NULL; + if (nf->frame == p->frame) { - nf->frame_index = ~0; + nf->frame = NULL; nf->flags &= ~VLIB_FRAME_IS_ALLOCATED; if (!(n->flags & VLIB_NODE_FLAG_FRAME_NO_FREE_AFTER_DISPATCH)) - restore_frame_index = p->frame_index; + restore_frame = p->frame; } /* Frame must be pending. */ - ASSERT (f->flags & VLIB_FRAME_PENDING); + ASSERT (f->frame_flags & VLIB_FRAME_PENDING); ASSERT (f->n_vectors > 0); /* Copy trace flag from next frame to node. @@ -1169,11 +1120,17 @@ dispatch_pending_node (vlib_main_t * vm, uword pending_frame_index, VLIB_NODE_TYPE_INTERNAL, VLIB_NODE_STATE_POLLING, f, last_time_stamp); + /* Internal node vector-rate accounting, for summary stats */ + vm->internal_node_vectors += f->n_vectors; + vm->internal_node_calls++; + vm->internal_node_last_vectors_per_main_loop = + (f->n_vectors > vm->internal_node_last_vectors_per_main_loop) ? + f->n_vectors : vm->internal_node_last_vectors_per_main_loop; - f->flags &= ~VLIB_FRAME_PENDING; + f->frame_flags &= ~(VLIB_FRAME_PENDING | VLIB_FRAME_NO_APPEND); /* Frame is ready to be used again, so restore it. */ - if (restore_frame_index != ~0) + if (restore_frame != NULL) { /* * We musn't restore a frame that is flagged to be freed. This @@ -1185,7 +1142,7 @@ dispatch_pending_node (vlib_main_t * vm, uword pending_frame_index, * longer part of the queue for that node and hence it cannot be * it's overspill. */ - ASSERT (!(f->flags & VLIB_FRAME_FREE_AFTER_DISPATCH)); + ASSERT (!(f->frame_flags & VLIB_FRAME_FREE_AFTER_DISPATCH)); /* * NB: dispatching node n can result in the creation and scheduling @@ -1201,26 +1158,27 @@ dispatch_pending_node (vlib_main_t * vm, uword pending_frame_index, nf = vec_elt_at_index (nm->next_frames, p->next_frame_index); nf->flags |= VLIB_FRAME_IS_ALLOCATED; - if (~0 == nf->frame_index) + if (NULL == nf->frame) { /* no new frame has been assigned to this node, use the saved one */ - nf->frame_index = restore_frame_index; + nf->frame = restore_frame; f->n_vectors = 0; + f->flags = 0; } else { /* The node has gained a frame, implying packets from the current frame were re-queued to this same node. we don't need the saved one anymore */ - vlib_frame_free (vm, n, f); + vlib_frame_free (vm, f); } } else { - if (f->flags & VLIB_FRAME_FREE_AFTER_DISPATCH) + if (f->frame_flags & VLIB_FRAME_FREE_AFTER_DISPATCH) { ASSERT (!(n->flags & VLIB_NODE_FLAG_FRAME_NO_FREE_AFTER_DISPATCH)); - vlib_frame_free (vm, n, f); + vlib_frame_free (vm, f); } } @@ -1255,6 +1213,8 @@ vlib_process_bootstrap (uword _a) vm = a->vm; p = a->process; + vlib_process_finish_switch_stack (vm); + f = a->frame; node = &p->node_runtime; @@ -1262,6 +1222,7 @@ vlib_process_bootstrap (uword _a) ASSERT (vlib_process_stack_is_valid (p)); + vlib_process_start_switch_stack (vm, 0); clib_longjmp (&p->return_longjmp, n); return n; @@ -1280,14 +1241,19 @@ vlib_process_startup (vlib_main_t * vm, vlib_process_t * p, vlib_frame_t * f) r = clib_setjmp (&p->return_longjmp, VLIB_PROCESS_RETURN_LONGJMP_RETURN); if (r == VLIB_PROCESS_RETURN_LONGJMP_RETURN) - r = clib_calljmp (vlib_process_bootstrap, pointer_to_uword (&a), - (void *) p->stack + (1 << p->log2_n_stack_bytes)); + { + vlib_process_start_switch_stack (vm, p); + r = clib_calljmp (vlib_process_bootstrap, pointer_to_uword (&a), + (void *) p->stack + (1 << p->log2_n_stack_bytes)); + } + else + vlib_process_finish_switch_stack (vm); return r; } static_always_inline uword -vlib_process_resume (vlib_process_t * p) +vlib_process_resume (vlib_main_t * vm, vlib_process_t * p) { uword r; p->flags &= ~(VLIB_PROCESS_IS_SUSPENDED_WAITING_FOR_CLOCK @@ -1295,7 +1261,12 @@ vlib_process_resume (vlib_process_t * p) | VLIB_PROCESS_RESUME_PENDING); r = clib_setjmp (&p->return_longjmp, VLIB_PROCESS_RETURN_LONGJMP_RETURN); if (r == VLIB_PROCESS_RETURN_LONGJMP_RETURN) - clib_longjmp (&p->resume_longjmp, VLIB_PROCESS_RESUME_LONGJMP_RESUME); + { + vlib_process_start_switch_stack (vm, p); + clib_longjmp (&p->resume_longjmp, VLIB_PROCESS_RESUME_LONGJMP_RESUME); + } + else + vlib_process_finish_switch_stack (vm); return r; } @@ -1306,6 +1277,7 @@ dispatch_process (vlib_main_t * vm, vlib_node_main_t *nm = &vm->node_main; vlib_node_runtime_t *node_runtime = &p->node_runtime; vlib_node_t *node = vlib_get_node (vm, node_runtime->node_index); + u32 old_process_index; u64 t; uword n_vectors, is_suspend; @@ -1321,11 +1293,15 @@ dispatch_process (vlib_main_t * vm, f ? f->n_vectors : 0, /* is_after */ 0); /* Save away current process for suspend. */ + old_process_index = nm->current_process_index; nm->current_process_index = node->runtime_index; + vlib_node_runtime_perf_counter (vm, node_runtime, f, 0, last_time_stamp, + VLIB_NODE_RUNTIME_PERF_BEFORE); + n_vectors = vlib_process_startup (vm, p, f); - nm->current_process_index = ~0; + nm->current_process_index = old_process_index; ASSERT (n_vectors != VLIB_PROCESS_RETURN_LONGJMP_RETURN); is_suspend = n_vectors == VLIB_PROCESS_RETURN_LONGJMP_SUSPEND; @@ -1336,7 +1312,7 @@ dispatch_process (vlib_main_t * vm, n_vectors = 0; pool_get (nm->suspended_process_frames, pf); pf->node_runtime_index = node->runtime_index; - pf->frame_index = f ? vlib_frame_index (vm, f) : ~0; + pf->frame = f; pf->next_frame_index = ~0; p->n_suspends += 1; @@ -1362,6 +1338,9 @@ dispatch_process (vlib_main_t * vm, vlib_elog_main_loop_event (vm, node_runtime->node_index, t, is_suspend, /* is_after */ 1); + vlib_node_runtime_perf_counter (vm, node_runtime, f, n_vectors, t, + VLIB_NODE_RUNTIME_PERF_AFTER); + vlib_process_update_stats (vm, p, /* n_calls */ !is_suspend, /* n_vectors */ n_vectors, @@ -1399,13 +1378,12 @@ dispatch_suspended_process (vlib_main_t * vm, ASSERT (p->flags & (VLIB_PROCESS_IS_SUSPENDED_WAITING_FOR_CLOCK | VLIB_PROCESS_IS_SUSPENDED_WAITING_FOR_EVENT)); - pf = - pool_elt_at_index (nm->suspended_process_frames, - p->suspended_process_frame_index); + pf = pool_elt_at_index (nm->suspended_process_frames, + p->suspended_process_frame_index); node_runtime = &p->node_runtime; node = vlib_get_node (vm, node_runtime->node_index); - f = pf->frame_index != ~0 ? vlib_get_frame (vm, pf->frame_index) : 0; + f = pf->frame; vlib_elog_main_loop_event (vm, node_runtime->node_index, t, f ? f->n_vectors : 0, /* is_after */ 0); @@ -1413,7 +1391,10 @@ dispatch_suspended_process (vlib_main_t * vm, /* Save away current process for suspend. */ nm->current_process_index = node->runtime_index; - n_vectors = vlib_process_resume (p); + vlib_node_runtime_perf_counter (vm, node_runtime, f, 0, last_time_stamp, + VLIB_NODE_RUNTIME_PERF_BEFORE); + + n_vectors = vlib_process_resume (vm, p); t = clib_cpu_time_now (); nm->current_process_index = ~0; @@ -1437,14 +1418,18 @@ dispatch_suspended_process (vlib_main_t * vm, else { p->flags &= ~VLIB_PROCESS_IS_RUNNING; + pool_put_index (nm->suspended_process_frames, + p->suspended_process_frame_index); p->suspended_process_frame_index = ~0; - pool_put (nm->suspended_process_frames, pf); } t = clib_cpu_time_now (); vlib_elog_main_loop_event (vm, node_runtime->node_index, t, !is_suspend, /* is_after */ 1); + vlib_node_runtime_perf_counter (vm, node_runtime, f, n_vectors, t, + VLIB_NODE_RUNTIME_PERF_AFTER); + vlib_process_update_stats (vm, p, /* n_calls */ !is_suspend, /* n_vectors */ n_vectors, @@ -1453,6 +1438,12 @@ dispatch_suspended_process (vlib_main_t * vm, return t; } +void vl_api_send_pending_rpc_requests (vlib_main_t *) __attribute__ ((weak)); +void +vl_api_send_pending_rpc_requests (vlib_main_t * vm) +{ +} + static_always_inline void vlib_main_or_worker_loop (vlib_main_t * vm, int is_main) { @@ -1460,14 +1451,15 @@ vlib_main_or_worker_loop (vlib_main_t * vm, int is_main) vlib_thread_main_t *tm = vlib_get_thread_main (); uword i; u64 cpu_time_now; + f64 now; vlib_frame_queue_main_t *fqm; - u32 *last_node_runtime_indices = 0; + u32 frame_queue_check_counter = 0; /* Initialize pending node vector. */ if (is_main) { vec_resize (nm->pending_frames, 32); - _vec_len (nm->pending_frames) = 0; + vec_set_len (nm->pending_frames, 0); } /* Mark time of main loop start. */ @@ -1480,10 +1472,7 @@ vlib_main_or_worker_loop (vlib_main_t * vm, int is_main) cpu_time_now = clib_cpu_time_now (); /* Pre-allocate interupt runtime indices and lock. */ - vec_alloc (nm->pending_interrupt_node_runtime_indices, 32); - vec_alloc (last_node_runtime_indices, 32); - if (!is_main) - clib_spinlock_init (&nm->pending_interrupt_lock); + vec_validate_aligned (nm->pending_interrupts, 0, CLIB_CACHE_LINE_BYTES); /* Pre-allocate expired nodes. */ if (!nm->polling_threshold_vector_length) @@ -1491,10 +1480,22 @@ vlib_main_or_worker_loop (vlib_main_t * vm, int is_main) if (!nm->interrupt_threshold_vector_length) nm->interrupt_threshold_vector_length = 5; + vm->cpu_id = clib_get_current_cpu_id (); + vm->numa_node = clib_get_current_numa_node (); + os_set_numa_index (vm->numa_node); + /* Start all processes. */ if (is_main) { uword i; + + /* + * Perform an initial barrier sync. Pays no attention to + * the barrier sync hold-down timer scheme, which won't work + * at this point in time. + */ + vlib_worker_thread_initial_barrier_sync_and_release (vm); + nm->current_process_index = ~0; for (i = 0; i < vec_len (nm->processes); i++) cpu_time_now = dispatch_process (vm, nm->processes[i], /* frame */ 0, @@ -1505,21 +1506,51 @@ vlib_main_or_worker_loop (vlib_main_t * vm, int is_main) { vlib_node_runtime_t *n; + if (PREDICT_FALSE (_vec_len (vm->pending_rpc_requests) > 0)) + { + if (!is_main) + vl_api_send_pending_rpc_requests (vm); + } + if (!is_main) + vlib_worker_thread_barrier_check (); + + if (PREDICT_FALSE (vm->check_frame_queues + frame_queue_check_counter)) { - vlib_worker_thread_barrier_check (); + u32 processed = 0; + vlib_frame_queue_dequeue_fn_t *fn; + + if (vm->check_frame_queues) + { + frame_queue_check_counter = 100; + vm->check_frame_queues = 0; + } + vec_foreach (fqm, tm->frame_queue_mains) - vlib_frame_queue_dequeue (vm, fqm); + { + fn = fqm->frame_queue_dequeue_fn; + processed += (fn) (vm, fqm); + } + + /* No handoff queue work found? */ + if (processed) + frame_queue_check_counter = 100; + else + frame_queue_check_counter--; } + if (PREDICT_FALSE (vec_len (vm->worker_thread_main_loop_callbacks))) + clib_call_callbacks (vm->worker_thread_main_loop_callbacks, vm, + cpu_time_now); + /* Process pre-input nodes. */ - if (is_main) - vec_foreach (n, nm->nodes_by_type[VLIB_NODE_TYPE_PRE_INPUT]) - cpu_time_now = dispatch_node (vm, n, - VLIB_NODE_TYPE_PRE_INPUT, - VLIB_NODE_STATE_POLLING, - /* frame */ 0, - cpu_time_now); + cpu_time_now = clib_cpu_time_now (); + vec_foreach (n, nm->nodes_by_type[VLIB_NODE_TYPE_PRE_INPUT]) + cpu_time_now = dispatch_node (vm, n, + VLIB_NODE_TYPE_PRE_INPUT, + VLIB_NODE_STATE_POLLING, + /* frame */ 0, + cpu_time_now); /* Next process input nodes. */ vec_foreach (n, nm->nodes_by_type[VLIB_NODE_TYPE_INPUT]) @@ -1532,53 +1563,75 @@ vlib_main_or_worker_loop (vlib_main_t * vm, int is_main) if (PREDICT_TRUE (is_main && vm->queue_signal_pending == 0)) vm->queue_signal_callback (vm); - /* Next handle interrupts. */ - { - uword l = _vec_len (nm->pending_interrupt_node_runtime_indices); - uword i; - if (l > 0) - { - u32 *tmp; - if (!is_main) - clib_spinlock_lock (&nm->pending_interrupt_lock); - tmp = nm->pending_interrupt_node_runtime_indices; - nm->pending_interrupt_node_runtime_indices = - last_node_runtime_indices; - last_node_runtime_indices = tmp; - _vec_len (last_node_runtime_indices) = 0; - if (!is_main) - clib_spinlock_unlock (&nm->pending_interrupt_lock); - for (i = 0; i < l; i++) - { - n = vec_elt_at_index (nm->nodes_by_type[VLIB_NODE_TYPE_INPUT], - last_node_runtime_indices[i]); - cpu_time_now = - dispatch_node (vm, n, VLIB_NODE_TYPE_INPUT, - VLIB_NODE_STATE_INTERRUPT, - /* frame */ 0, - cpu_time_now); - } - } - } + if (__atomic_load_n (nm->pending_interrupts, __ATOMIC_ACQUIRE)) + { + int int_num = -1; + *nm->pending_interrupts = 0; + + while ((int_num = + clib_interrupt_get_next (nm->interrupts, int_num)) != -1) + { + vlib_node_runtime_t *n; + clib_interrupt_clear (nm->interrupts, int_num); + n = vec_elt_at_index (nm->nodes_by_type[VLIB_NODE_TYPE_INPUT], + int_num); + cpu_time_now = dispatch_node (vm, n, VLIB_NODE_TYPE_INPUT, + VLIB_NODE_STATE_INTERRUPT, + /* frame */ 0, cpu_time_now); + } + } + + /* Input nodes may have added work to the pending vector. + Process pending vector until there is nothing left. + All pending vectors will be processed from input -> output. */ + for (i = 0; i < _vec_len (nm->pending_frames); i++) + cpu_time_now = dispatch_pending_node (vm, i, cpu_time_now); + /* Reset pending vector for next iteration. */ + vec_set_len (nm->pending_frames, 0); if (is_main) { + /* *INDENT-OFF* */ + ELOG_TYPE_DECLARE (es) = + { + .format = "process tw start", + .format_args = "", + }; + ELOG_TYPE_DECLARE (ee) = + { + .format = "process tw end: %d", + .format_args = "i4", + }; + /* *INDENT-ON* */ + + struct + { + int nready_procs; + } *ed; + /* Check if process nodes have expired from timing wheel. */ ASSERT (nm->data_from_advancing_timing_wheel != 0); - nm->data_from_advancing_timing_wheel = - TW (tw_timer_expire_timers_vec) - ((TWT (tw_timer_wheel) *) nm->timing_wheel, vlib_time_now (vm), - nm->data_from_advancing_timing_wheel); + if (PREDICT_FALSE (vm->elog_trace_graph_dispatch)) + ed = ELOG_DATA (&vlib_global_main.elog_main, es); + + TW (tw_timer_expire_timers) + ((TWT (tw_timer_wheel) *) nm->timing_wheel, vlib_time_now (vm)); ASSERT (nm->data_from_advancing_timing_wheel != 0); + if (PREDICT_FALSE (vm->elog_trace_graph_dispatch)) + { + ed = ELOG_DATA (&vlib_global_main.elog_main, ee); + ed->nready_procs = + _vec_len (nm->data_from_advancing_timing_wheel); + } + if (PREDICT_FALSE (_vec_len (nm->data_from_advancing_timing_wheel) > 0)) { uword i; - processes_timing_wheel_data: for (i = 0; i < _vec_len (nm->data_from_advancing_timing_wheel); i++) { @@ -1601,12 +1654,12 @@ vlib_main_or_worker_loop (vlib_main_t * vm, int is_main) te->n_data_elts, te->n_data_elt_bytes); if (te->n_data_bytes < sizeof (te->inline_event_data)) - clib_memcpy (data, te->inline_event_data, - te->n_data_bytes); + clib_memcpy_fast (data, te->inline_event_data, + te->n_data_bytes); else { - clib_memcpy (data, te->event_data_as_vector, - te->n_data_bytes); + clib_memcpy_fast (data, te->event_data_as_vector, + te->n_data_bytes); vec_free (te->event_data_as_vector); } pool_put (nm->signal_timed_event_data_pool, te); @@ -1618,27 +1671,40 @@ vlib_main_or_worker_loop (vlib_main_t * vm, int is_main) dispatch_suspended_process (vm, di, cpu_time_now); } } - _vec_len (nm->data_from_advancing_timing_wheel) = 0; + vec_set_len (nm->data_from_advancing_timing_wheel, 0); } } - - /* Input nodes may have added work to the pending vector. - Process pending vector until there is nothing left. - All pending vectors will be processed from input -> output. */ - for (i = 0; i < _vec_len (nm->pending_frames); i++) - cpu_time_now = dispatch_pending_node (vm, i, cpu_time_now); - /* Reset pending vector for next iteration. */ - _vec_len (nm->pending_frames) = 0; - - /* Pending internal nodes may resume processes. */ - if (is_main && _vec_len (nm->data_from_advancing_timing_wheel) > 0) - goto processes_timing_wheel_data; - vlib_increment_main_loop_counter (vm); - /* Record time stamp in case there are no enabled nodes and above calls do not update time stamp. */ cpu_time_now = clib_cpu_time_now (); + vm->loops_this_reporting_interval++; + now = clib_time_now_internal (&vm->clib_time, cpu_time_now); + /* Time to update loops_per_second? */ + if (PREDICT_FALSE (now >= vm->loop_interval_end)) + { + /* Next sample ends in 20ms */ + if (vm->loop_interval_start) + { + f64 this_loops_per_second; + + this_loops_per_second = + ((f64) vm->loops_this_reporting_interval) / (now - + vm->loop_interval_start); + + vm->loops_per_second = + vm->loops_per_second * vm->damping_constant + + (1.0 - vm->damping_constant) * this_loops_per_second; + if (vm->loops_per_second != 0.0) + vm->seconds_per_loop = 1.0 / vm->loops_per_second; + else + vm->seconds_per_loop = 0.0; + } + /* New interval starts now, and ends in 20ms */ + vm->loop_interval_start = now; + vm->loop_interval_end = now + 2e-4; + vm->loops_this_reporting_interval = 0; + } } } @@ -1654,11 +1720,49 @@ vlib_worker_loop (vlib_main_t * vm) vlib_main_or_worker_loop (vm, /* is_main */ 0); } -vlib_main_t vlib_global_main; +vlib_global_main_t vlib_global_main; + +void +vlib_add_del_post_mortem_callback (void *cb, int is_add) +{ + vlib_global_main_t *vgm = vlib_get_global_main (); + int i; + + if (is_add == 0) + { + for (i = vec_len (vgm->post_mortem_callbacks) - 1; i >= 0; i--) + if (vgm->post_mortem_callbacks[i] == cb) + vec_del1 (vgm->post_mortem_callbacks, i); + return; + } + + for (i = 0; i < vec_len (vgm->post_mortem_callbacks); i++) + if (vgm->post_mortem_callbacks[i] == cb) + return; + vec_add1 (vgm->post_mortem_callbacks, cb); +} + +static void +elog_post_mortem_dump (void) +{ + elog_main_t *em = vlib_get_elog_main (); + + u8 *filename; + clib_error_t *error; + + filename = format (0, "/tmp/elog_post_mortem.%d%c", getpid (), 0); + error = elog_write_file (em, (char *) filename, 1 /* flush ring */); + if (error) + clib_error_report (error); + /* + * We're in the middle of crashing. Don't try to free the filename. + */ +} static clib_error_t * vlib_main_configure (vlib_main_t * vm, unformat_input_t * input) { + vlib_global_main_t *vgm = vlib_get_global_main (); int turn_on_mem_trace = 0; while (unformat_check_input (input) != UNFORMAT_END_OF_INPUT) @@ -1667,10 +1771,26 @@ vlib_main_configure (vlib_main_t * vm, unformat_input_t * input) turn_on_mem_trace = 1; else if (unformat (input, "elog-events %d", - &vm->elog_main.event_ring_size)) - ; + &vgm->configured_elog_ring_size)) + vgm->configured_elog_ring_size = + 1 << max_log2 (vgm->configured_elog_ring_size); else if (unformat (input, "elog-post-mortem-dump")) - vm->elog_post_mortem_dump = 1; + vlib_add_del_post_mortem_callback (elog_post_mortem_dump, + /* is_add */ 1); + else if (unformat (input, "buffer-alloc-success-rate %f", + &vm->buffer_alloc_success_rate)) + { + if (VLIB_BUFFER_ALLOC_FAULT_INJECTOR == 0) + return clib_error_return + (0, "Buffer fault injection not configured"); + } + else if (unformat (input, "buffer-alloc-success-seed %u", + &vm->buffer_alloc_success_seed)) + { + if (VLIB_BUFFER_ALLOC_FAULT_INJECTOR == 0) + return clib_error_return + (0, "Buffer fault injection not configured"); + } else return unformat_parse_error (input); } @@ -1687,47 +1807,104 @@ vlib_main_configure (vlib_main_t * vm, unformat_input_t * input) VLIB_EARLY_CONFIG_FUNCTION (vlib_main_configure, "vlib"); static void -dummy_queue_signal_callback (vlib_main_t * vm) +placeholder_queue_signal_callback (vlib_main_t * vm) +{ +} + +#define foreach_weak_reference_stub \ +_(vpe_api_init) \ +_(vlibmemory_init) \ +_(map_api_segment_init) + +#define _(name) \ +clib_error_t *name (vlib_main_t *vm) __attribute__((weak)); \ +clib_error_t *name (vlib_main_t *vm) { return 0; } +foreach_weak_reference_stub; +#undef _ + +void vl_api_set_elog_main (elog_main_t * m) __attribute__ ((weak)); +void +vl_api_set_elog_main (elog_main_t * m) +{ + clib_warning ("STUB"); +} + +int vl_api_set_elog_trace_api_messages (int enable) __attribute__ ((weak)); +int +vl_api_set_elog_trace_api_messages (int enable) +{ + clib_warning ("STUB"); + return 0; +} + +int vl_api_get_elog_trace_api_messages (void) __attribute__ ((weak)); +int +vl_api_get_elog_trace_api_messages (void) +{ + clib_warning ("STUB"); + return 0; +} + +static void +process_expired_timer_cb (u32 *expired_timer_handles) { + vlib_main_t *vm = vlib_get_main (); + vlib_node_main_t *nm = &vm->node_main; + u32 *handle; + + vec_foreach (handle, expired_timer_handles) + { + u32 pi = vlib_timing_wheel_data_get_index (*handle); + vlib_process_t *p = vec_elt (nm->processes, pi); + + p->stop_timer_handle = ~0; + } + vec_append (nm->data_from_advancing_timing_wheel, expired_timer_handles); } /* Main function. */ int vlib_main (vlib_main_t * volatile vm, unformat_input_t * input) { + vlib_global_main_t *vgm = vlib_get_global_main (); clib_error_t *volatile error; vlib_node_main_t *nm = &vm->node_main; - vm->queue_signal_callback = dummy_queue_signal_callback; - - clib_time_init (&vm->clib_time); + vm->queue_signal_callback = placeholder_queue_signal_callback; - /* Turn on event log. */ - if (!vm->elog_main.event_ring_size) - vm->elog_main.event_ring_size = 128 << 10; - elog_init (&vm->elog_main, vm->elog_main.event_ring_size); - elog_enable_disable (&vm->elog_main, 1); + /* Reconfigure event log which is enabled very early */ + if (vgm->configured_elog_ring_size && + vgm->configured_elog_ring_size != vgm->elog_main.event_ring_size) + elog_resize (&vgm->elog_main, vgm->configured_elog_ring_size); + vl_api_set_elog_main (vlib_get_elog_main ()); + (void) vl_api_set_elog_trace_api_messages (1); /* Default name. */ - if (!vm->name) - vm->name = "VLIB"; + if (!vgm->name) + vgm->name = "VLIB"; - vec_validate (vm->buffer_main, 0); - if (vlib_buffer_callbacks) + if ((error = vlib_physmem_init (vm))) { - /* external plugin has registered own buffer callbacks - so we just copy them */ - vlib_buffer_main_t *bm = vm->buffer_main; - clib_memcpy (&bm->cb, vlib_buffer_callbacks, - sizeof (vlib_buffer_callbacks_t)); - bm->callbacks_registered = 1; + clib_error_report (error); + goto done; } - else + + if ((error = vlib_log_init (vm))) { - vlib_physmem_main_t *vpm = &vm->physmem_main; - vlib_buffer_cb_init (vm); - unix_physmem_init (vm, 0 /* fail_if_physical_memory_not_present */ ); - vlib_buffer_add_mem_range (vm, vpm->virtual.start, vpm->virtual.size); + clib_error_report (error); + goto done; + } + + if ((error = vlib_stats_init (vm))) + { + clib_error_report (error); + goto done; + } + + if ((error = vlib_buffer_main_init (vm))) + { + clib_error_report (error); + goto done; } if ((error = vlib_thread_init (vm))) @@ -1736,6 +1913,9 @@ vlib_main (vlib_main_t * volatile vm, unformat_input_t * input) goto done; } + /* Register node ifunction variants */ + vlib_register_all_node_march_variants (vm); + /* Register static nodes so that init functions may use them. */ vlib_register_all_static_nodes (vm); @@ -1755,28 +1935,78 @@ vlib_main (vlib_main_t * volatile vm, unformat_input_t * input) goto done; } + /* Direct call / weak reference, for vlib standalone use-cases */ + if ((error = vpe_api_init (vm))) + { + clib_error_report (error); + goto done; + } + + if ((error = vlibmemory_init (vm))) + { + clib_error_report (error); + goto done; + } + + if ((error = map_api_segment_init (vm))) + { + clib_error_report (error); + goto done; + } + /* See unix/main.c; most likely already set up */ - if (vm->init_functions_called == 0) - vm->init_functions_called = hash_create (0, /* value bytes */ 0); + if (vgm->init_functions_called == 0) + vgm->init_functions_called = hash_create (0, /* value bytes */ 0); if ((error = vlib_call_all_init_functions (vm))) goto done; - /* Create default buffer free list. */ - vlib_buffer_get_or_create_free_list (vm, - VLIB_BUFFER_DEFAULT_FREE_LIST_BYTES, - "default"); - nm->timing_wheel = clib_mem_alloc_aligned (sizeof (TWT (tw_timer_wheel)), CLIB_CACHE_LINE_BYTES); vec_validate (nm->data_from_advancing_timing_wheel, 10); - _vec_len (nm->data_from_advancing_timing_wheel) = 0; + vec_set_len (nm->data_from_advancing_timing_wheel, 0); /* Create the process timing wheel */ - TW (tw_timer_wheel_init) ((TWT (tw_timer_wheel) *) nm->timing_wheel, - 0 /* no callback */ , - 10e-6 /* timer period 10us */ , - ~0 /* max expirations per call */ ); + TW (tw_timer_wheel_init) + ((TWT (tw_timer_wheel) *) nm->timing_wheel, + process_expired_timer_cb /* callback */, 10e-6 /* timer period 10us */, + ~0 /* max expirations per call */); + + vec_validate (vm->pending_rpc_requests, 0); + vec_set_len (vm->pending_rpc_requests, 0); + vec_validate (vm->processing_rpc_requests, 0); + vec_set_len (vm->processing_rpc_requests, 0); + + /* Default params for the buffer allocator fault injector, if configured */ + if (VLIB_BUFFER_ALLOC_FAULT_INJECTOR > 0) + { + vm->buffer_alloc_success_seed = 0xdeaddabe; + vm->buffer_alloc_success_rate = 0.80; + } + + if ((error = vlib_call_all_config_functions (vm, input, 0 /* is_early */ ))) + goto done; + + /* + * Use exponential smoothing, with a half-life of 1 second + * reported_rate(t) = reported_rate(t-1) * K + rate(t)*(1-K) + * + * Sample every 20ms, aka 50 samples per second + * K = exp (-1.0/20.0); + * K = 0.95 + */ + vm->damping_constant = exp (-1.0 / 20.0); + + /* Sort per-thread init functions before we start threads */ + vlib_sort_init_exit_functions (&vgm->worker_init_function_registrations); + + /* Call all main loop enter functions. */ + { + clib_error_t *sub_error; + sub_error = vlib_call_all_main_loop_enter_functions (vm); + if (sub_error) + clib_error_report (sub_error); + } switch (clib_setjmp (&vm->main_loop_exit, VLIB_MAIN_LOOP_EXIT_NONE)) { @@ -1792,20 +2022,12 @@ vlib_main (vlib_main_t * volatile vm, unformat_input_t * input) goto done; } - if ((error = vlib_call_all_config_functions (vm, input, 0 /* is_early */ ))) - goto done; - - /* Call all main loop enter functions. */ - { - clib_error_t *sub_error; - sub_error = vlib_call_all_main_loop_enter_functions (vm); - if (sub_error) - clib_error_report (sub_error); - } - vlib_main_loop (vm); done: + /* Stop worker threads, barrier will not be released */ + vlib_worker_thread_barrier_sync (vm); + /* Call all exit functions. */ { clib_error_t *sub_error; @@ -1817,7 +2039,26 @@ done: if (error) clib_error_report (error); - return 0; + return vm->main_loop_exit_status; +} + +vlib_main_t * +vlib_get_main_not_inline (void) +{ + return vlib_get_main (); +} + +elog_main_t * +vlib_get_elog_main_not_inline () +{ + return &vlib_global_main.elog_main; +} + +void +vlib_exit_with_status (vlib_main_t *vm, int status) +{ + vm->main_loop_exit_status = status; + __atomic_store_n (&vm->main_loop_exit_now, 1, __ATOMIC_RELEASE); } /*