#include <asm/unistd.h>
#include <sys/ioctl.h>
+/* "not in glibc" */
static long
perf_event_open (struct perf_event_attr *hw_event, pid_t pid, int cpu,
int group_fd, unsigned long flags)
}
static void
-read_current_perf_counters (vlib_main_t * vm, u64 * c0, u64 * c1)
+read_current_perf_counters (vlib_main_t * vm, u64 * c0, u64 * c1,
+ vlib_node_runtime_t * node,
+ vlib_frame_t * frame, int before_or_after)
{
int i;
u64 *cc;
else
{
u64 sw_value;
- if (read (pm->pm_fds[i][my_thread_index], &sw_value,
- sizeof (sw_value)) != sizeof (sw_value))
+ int read_result;
+ if ((read_result = read (pm->pm_fds[i][my_thread_index], &sw_value,
+ sizeof (sw_value)) != sizeof (sw_value)))
{
clib_unix_warning
- ("counter read failed, disable collection...");
- vm->vlib_node_runtime_perf_counter_cb = 0;
+ ("counter read returned %d, expected %d",
+ read_result, sizeof (sw_value));
+ clib_callback_enable_disable
+ (vm->vlib_node_runtime_perf_counter_cbs,
+ vm->vlib_node_runtime_perf_counter_cb_tmp,
+ vm->worker_thread_main_loop_callback_lock,
+ read_current_perf_counters, 0 /* enable */ );
return;
}
*cc = sw_value;
u32 my_thread_index = vm->thread_index;
u32 index;
int i, limit = 1;
+ int cpu;
if ((pm->current_event + 1) < vec_len (pm->single_events_to_collect))
limit = 2;
for (i = 0; i < limit; i++)
{
+ vec_validate (pm->pm_fds[i], vec_len (vlib_mains) - 1);
+ vec_validate (pm->perf_event_pages[i], vec_len (vlib_mains) - 1);
+ vec_validate (pm->rdpmc_indices[i], vec_len (vlib_mains) - 1);
+
c = vec_elt_at_index (pm->single_events_to_collect,
pm->current_event + i);
pe.exclude_hv = 1;
}
- fd = perf_event_open (&pe, 0, -1, -1, 0);
+ cpu = vm->cpu_id;
+
+ fd = perf_event_open (&pe, 0, cpu, -1, 0);
if (fd == -1)
{
clib_unix_warning ("event open: type %d config %d", c->pe_type,
close (fd);
return;
}
+ CLIB_MEM_UNPOISON (p, pm->page_size);
}
else
p = 0;
if (ioctl (fd, PERF_EVENT_IOC_ENABLE, 0) < 0)
clib_unix_warning ("enable ioctl");
+ pm->perf_event_pages[i][my_thread_index] = (void *) p;
+ pm->pm_fds[i][my_thread_index] = fd;
+ }
+
+ /*
+ * Hardware events must be all opened and enabled before aquiring
+ * pmc indices, otherwise the pmc indices might be out-dated.
+ */
+ for (i = 0; i < limit; i++)
+ {
+ p =
+ (struct perf_event_mmap_page *)
+ pm->perf_event_pages[i][my_thread_index];
+
/*
* Software event counters - and others not capable of being
* read via the "rdpmc" instruction - will be read
* by system calls.
*/
- if (pe.type == PERF_TYPE_SOFTWARE || p->cap_user_rdpmc == 0)
+ if (p == 0 || p->cap_user_rdpmc == 0)
index = ~0;
else
index = p->index - 1;
pm->rdpmc_indices[i][my_thread_index] = index;
- pm->perf_event_pages[i][my_thread_index] = (void *) p;
- pm->pm_fds[i][my_thread_index] = fd;
}
pm->n_active = i;
/* Enable the main loop counter snapshot mechanism */
- vm->vlib_node_runtime_perf_counter_cb = read_current_perf_counters;
+ clib_callback_enable_disable
+ (vm->vlib_node_runtime_perf_counter_cbs,
+ vm->vlib_node_runtime_perf_counter_cb_tmp,
+ vm->worker_thread_main_loop_callback_lock,
+ read_current_perf_counters, 1 /* enable */ );
}
static void
int i;
/* Stop main loop collection */
- vm->vlib_node_runtime_perf_counter_cb = 0;
+ clib_callback_enable_disable
+ (vm->vlib_node_runtime_perf_counter_cbs,
+ vm->vlib_node_runtime_perf_counter_cb_tmp,
+ vm->worker_thread_main_loop_callback_lock,
+ read_current_perf_counters, 0 /* enable */ );
for (i = 0; i < pm->n_active; i++)
{
clib_unix_warning ("disable ioctl");
if (pm->perf_event_pages[i][my_thread_index])
- if (munmap (pm->perf_event_pages[i][my_thread_index],
- pm->page_size) < 0)
- clib_unix_warning ("munmap");
+ {
+ if (munmap (pm->perf_event_pages[i][my_thread_index],
+ pm->page_size) < 0)
+ clib_unix_warning ("munmap");
+ CLIB_MEM_POISON (pm->perf_event_pages[i][my_thread_index],
+ pm->page_size);
+ pm->perf_event_pages[i][my_thread_index] = 0;
+ }
(void) close (pm->pm_fds[i][my_thread_index]);
pm->pm_fds[i][my_thread_index] = 0;
+
}
}
{
perfmon_main_t *pm = &perfmon_main;
+ clib_callback_enable_disable (vm->worker_thread_main_loop_callbacks,
+ vm->worker_thread_main_loop_callback_tmp,
+ vm->worker_thread_main_loop_callback_lock,
+ worker_thread_start_event, 0 /* enable */ );
enable_current_events (pm);
- vm->worker_thread_main_loop_callback = 0;
}
static void
worker_thread_stop_event (vlib_main_t * vm)
{
perfmon_main_t *pm = &perfmon_main;
+ clib_callback_enable_disable (vm->worker_thread_main_loop_callbacks,
+ vm->worker_thread_main_loop_callback_tmp,
+ vm->worker_thread_main_loop_callback_lock,
+ worker_thread_stop_event, 0 /* enable */ );
disable_events (pm);
- vm->worker_thread_main_loop_callback = 0;
}
static void
start_event (perfmon_main_t * pm, f64 now, uword event_data)
{
int i;
+ int last_set;
+ int all = 0;
pm->current_event = 0;
+
if (vec_len (pm->single_events_to_collect) == 0)
{
pm->state = PERFMON_STATE_OFF;
return;
}
+
+ last_set = clib_bitmap_last_set (pm->thread_bitmap);
+ all = (last_set == ~0);
+
pm->state = PERFMON_STATE_RUNNING;
clear_counters (pm);
- /* Start collection on this thread */
- enable_current_events (pm);
+ /* Start collection on thread 0? */
+ if (all || clib_bitmap_get (pm->thread_bitmap, 0))
+ {
+ /* Start collection on this thread */
+ enable_current_events (pm);
+ }
/* And also on worker threads */
for (i = 1; i < vec_len (vlib_mains); i++)
{
if (vlib_mains[i] == 0)
continue;
- vlib_mains[i]->worker_thread_main_loop_callback = (void *)
- worker_thread_start_event;
+
+ if (all || clib_bitmap_get (pm->thread_bitmap, i))
+ clib_callback_enable_disable
+ (vlib_mains[i]->worker_thread_main_loop_callbacks,
+ vlib_mains[i]->worker_thread_main_loop_callback_tmp,
+ vlib_mains[i]->worker_thread_main_loop_callback_lock,
+ (void *) worker_thread_start_event, 1 /* enable */ );
}
}
}
static void
-handle_timeout (perfmon_main_t * pm, f64 now)
+handle_timeout (vlib_main_t * vm, perfmon_main_t * pm, f64 now)
{
int i;
- disable_events (pm);
+ int last_set, all;
+
+ last_set = clib_bitmap_last_set (pm->thread_bitmap);
+ all = (last_set == ~0);
+
+ if (all || clib_bitmap_get (pm->thread_bitmap, 0))
+ disable_events (pm);
/* And also on worker threads */
for (i = 1; i < vec_len (vlib_mains); i++)
{
if (vlib_mains[i] == 0)
continue;
- vlib_mains[i]->worker_thread_main_loop_callback = (void *)
- worker_thread_stop_event;
+ if (all || clib_bitmap_get (pm->thread_bitmap, i))
+ clib_callback_enable_disable
+ (vlib_mains[i]->worker_thread_main_loop_callbacks,
+ vlib_mains[i]->worker_thread_main_loop_callback_tmp,
+ vlib_mains[i]->worker_thread_main_loop_callback_lock,
+ (void *) worker_thread_stop_event, 1 /* enable */ );
}
- /* Short delay to make sure workers have stopped collection */
+ /* Make sure workers have stopped collection */
if (i > 1)
- vlib_process_suspend (pm->vlib_main, 1e-3);
+ {
+ f64 deadman = vlib_time_now (vm) + 1.0;
+
+ for (i = 1; i < vec_len (vlib_mains); i++)
+ {
+ /* Has the worker actually stopped collecting data? */
+ while (clib_callback_is_set
+ (vlib_mains[i]->worker_thread_main_loop_callbacks,
+ vlib_mains[i]->worker_thread_main_loop_callback_lock,
+ read_current_perf_counters))
+ {
+ if (vlib_time_now (vm) > deadman)
+ {
+ clib_warning ("Thread %d deadman timeout!", i);
+ break;
+ }
+ vlib_process_suspend (pm->vlib_main, 1e-3);
+ }
+ }
+ }
scrape_and_clear_counters (pm);
pm->current_event += pm->n_active;
if (pm->current_event >= vec_len (pm->single_events_to_collect))
pm->state = PERFMON_STATE_OFF;
return;
}
- enable_current_events (pm);
+
+ if (all || clib_bitmap_get (pm->thread_bitmap, 0))
+ enable_current_events (pm);
/* And also on worker threads */
for (i = 1; i < vec_len (vlib_mains); i++)
{
if (vlib_mains[i] == 0)
continue;
- vlib_mains[i]->worker_thread_main_loop_callback = (void *)
- worker_thread_start_event;
+ if (all || clib_bitmap_get (pm->thread_bitmap, i))
+ clib_callback_enable_disable
+ (vlib_mains[i]->worker_thread_main_loop_callbacks,
+ vlib_mains[i]->worker_thread_main_loop_callback_tmp,
+ vlib_mains[i]->worker_thread_main_loop_callback_lock,
+ worker_thread_start_event, 1 /* enable */ );
}
}
/* Handle timeout */
case ~0:
- handle_timeout (pm, now);
+ handle_timeout (vm, pm, now);
break;
default: