misc: asan: do not poison memory after munmap()
[vpp.git] / src / plugins / perfmon / perfmon_periodic.c
index ae20ac4..37d669b 100644 (file)
@@ -21,6 +21,7 @@
 #include <asm/unistd.h>
 #include <sys/ioctl.h>
 
+/* "not in glibc" */
 static long
 perf_event_open (struct perf_event_attr *hw_event, pid_t pid, int cpu,
                 int group_fd, unsigned long flags)
@@ -32,7 +33,9 @@ perf_event_open (struct perf_event_attr *hw_event, pid_t pid, int cpu,
 }
 
 static void
-read_current_perf_counters (vlib_main_t * vm, u64 * c0, u64 * c1)
+read_current_perf_counters (vlib_main_t * vm, u64 * c0, u64 * c1,
+                           vlib_node_runtime_t * node,
+                           vlib_frame_t * frame, int before_or_after)
 {
   int i;
   u64 *cc;
@@ -49,12 +52,18 @@ read_current_perf_counters (vlib_main_t * vm, u64 * c0, u64 * c1)
       else
        {
          u64 sw_value;
-         if (read (pm->pm_fds[i][my_thread_index], &sw_value,
-                   sizeof (sw_value)) != sizeof (sw_value))
+         int read_result;
+         if ((read_result = read (pm->pm_fds[i][my_thread_index], &sw_value,
+                                  sizeof (sw_value)) != sizeof (sw_value)))
            {
              clib_unix_warning
-               ("counter read failed, disable collection...");
-             vm->vlib_node_runtime_perf_counter_cb = 0;
+               ("counter read returned %d, expected %d",
+                read_result, sizeof (sw_value));
+             clib_callback_enable_disable
+               (vm->vlib_node_runtime_perf_counter_cbs,
+                vm->vlib_node_runtime_perf_counter_cb_tmp,
+                vm->worker_thread_main_loop_callback_lock,
+                read_current_perf_counters, 0 /* enable */ );
              return;
            }
          *cc = sw_value;
@@ -114,12 +123,17 @@ enable_current_events (perfmon_main_t * pm)
   u32 my_thread_index = vm->thread_index;
   u32 index;
   int i, limit = 1;
+  int cpu;
 
   if ((pm->current_event + 1) < vec_len (pm->single_events_to_collect))
     limit = 2;
 
   for (i = 0; i < limit; i++)
     {
+      vec_validate (pm->pm_fds[i], vec_len (vlib_mains) - 1);
+      vec_validate (pm->perf_event_pages[i], vec_len (vlib_mains) - 1);
+      vec_validate (pm->rdpmc_indices[i], vec_len (vlib_mains) - 1);
+
       c = vec_elt_at_index (pm->single_events_to_collect,
                            pm->current_event + i);
 
@@ -140,7 +154,9 @@ enable_current_events (perfmon_main_t * pm)
          pe.exclude_hv = 1;
        }
 
-      fd = perf_event_open (&pe, 0, -1, -1, 0);
+      cpu = vm->cpu_id;
+
+      fd = perf_event_open (&pe, 0, cpu, -1, 0);
       if (fd == -1)
        {
          clib_unix_warning ("event open: type %d config %d", c->pe_type,
@@ -157,34 +173,51 @@ enable_current_events (perfmon_main_t * pm)
              close (fd);
              return;
            }
+         CLIB_MEM_UNPOISON (p, pm->page_size);
        }
       else
        p = 0;
 
+      if (ioctl (fd, PERF_EVENT_IOC_RESET, 0) < 0)
+       clib_unix_warning ("reset ioctl");
+
+      if (ioctl (fd, PERF_EVENT_IOC_ENABLE, 0) < 0)
+       clib_unix_warning ("enable ioctl");
+
+      pm->perf_event_pages[i][my_thread_index] = (void *) p;
+      pm->pm_fds[i][my_thread_index] = fd;
+    }
+
+  /*
+   * Hardware events must be all opened and enabled before aquiring
+   * pmc indices, otherwise the pmc indices might be out-dated.
+   */
+  for (i = 0; i < limit; i++)
+    {
+      p =
+       (struct perf_event_mmap_page *)
+       pm->perf_event_pages[i][my_thread_index];
+
       /*
        * Software event counters - and others not capable of being
        * read via the "rdpmc" instruction - will be read
        * by system calls.
        */
-      if (pe.type == PERF_TYPE_SOFTWARE || p->cap_user_rdpmc == 0)
+      if (p == 0 || p->cap_user_rdpmc == 0)
        index = ~0;
       else
        index = p->index - 1;
 
-      if (ioctl (fd, PERF_EVENT_IOC_RESET, 0) < 0)
-       clib_unix_warning ("reset ioctl");
-
-      if (ioctl (fd, PERF_EVENT_IOC_ENABLE, 0) < 0)
-       clib_unix_warning ("enable ioctl");
-
       pm->rdpmc_indices[i][my_thread_index] = index;
-      pm->perf_event_pages[i][my_thread_index] = (void *) p;
-      pm->pm_fds[i][my_thread_index] = fd;
     }
 
   pm->n_active = i;
   /* Enable the main loop counter snapshot mechanism */
-  vm->vlib_node_runtime_perf_counter_cb = read_current_perf_counters;
+  clib_callback_enable_disable
+    (vm->vlib_node_runtime_perf_counter_cbs,
+     vm->vlib_node_runtime_perf_counter_cb_tmp,
+     vm->worker_thread_main_loop_callback_lock,
+     read_current_perf_counters, 1 /* enable */ );
 }
 
 static void
@@ -195,7 +228,11 @@ disable_events (perfmon_main_t * pm)
   int i;
 
   /* Stop main loop collection */
-  vm->vlib_node_runtime_perf_counter_cb = 0;
+  clib_callback_enable_disable
+    (vm->vlib_node_runtime_perf_counter_cbs,
+     vm->vlib_node_runtime_perf_counter_cb_tmp,
+     vm->worker_thread_main_loop_callback_lock,
+     read_current_perf_counters, 0 /* enable */ );
 
   for (i = 0; i < pm->n_active; i++)
     {
@@ -207,12 +244,16 @@ disable_events (perfmon_main_t * pm)
        clib_unix_warning ("disable ioctl");
 
       if (pm->perf_event_pages[i][my_thread_index])
-       if (munmap (pm->perf_event_pages[i][my_thread_index],
-                   pm->page_size) < 0)
-         clib_unix_warning ("munmap");
+       {
+         if (munmap (pm->perf_event_pages[i][my_thread_index],
+                     pm->page_size) < 0)
+           clib_unix_warning ("munmap");
+         pm->perf_event_pages[i][my_thread_index] = 0;
+       }
 
       (void) close (pm->pm_fds[i][my_thread_index]);
       pm->pm_fds[i][my_thread_index] = 0;
+
     }
 }
 
@@ -221,41 +262,63 @@ worker_thread_start_event (vlib_main_t * vm)
 {
   perfmon_main_t *pm = &perfmon_main;
 
+  clib_callback_enable_disable (vm->worker_thread_main_loop_callbacks,
+                               vm->worker_thread_main_loop_callback_tmp,
+                               vm->worker_thread_main_loop_callback_lock,
+                               worker_thread_start_event, 0 /* enable */ );
   enable_current_events (pm);
-  vm->worker_thread_main_loop_callback = 0;
 }
 
 static void
 worker_thread_stop_event (vlib_main_t * vm)
 {
   perfmon_main_t *pm = &perfmon_main;
+  clib_callback_enable_disable (vm->worker_thread_main_loop_callbacks,
+                               vm->worker_thread_main_loop_callback_tmp,
+                               vm->worker_thread_main_loop_callback_lock,
+                               worker_thread_stop_event, 0 /* enable */ );
   disable_events (pm);
-  vm->worker_thread_main_loop_callback = 0;
 }
 
 static void
 start_event (perfmon_main_t * pm, f64 now, uword event_data)
 {
   int i;
+  int last_set;
+  int all = 0;
   pm->current_event = 0;
+
   if (vec_len (pm->single_events_to_collect) == 0)
     {
       pm->state = PERFMON_STATE_OFF;
       return;
     }
+
+  last_set = clib_bitmap_last_set (pm->thread_bitmap);
+  all = (last_set == ~0);
+
   pm->state = PERFMON_STATE_RUNNING;
   clear_counters (pm);
 
-  /* Start collection on this thread */
-  enable_current_events (pm);
+  /* Start collection on thread 0? */
+  if (all || clib_bitmap_get (pm->thread_bitmap, 0))
+    {
+      /* Start collection on this thread */
+      enable_current_events (pm);
+    }
 
   /* And also on worker threads */
   for (i = 1; i < vec_len (vlib_mains); i++)
     {
       if (vlib_mains[i] == 0)
        continue;
-      vlib_mains[i]->worker_thread_main_loop_callback = (void *)
-       worker_thread_start_event;
+
+      if (all || clib_bitmap_get (pm->thread_bitmap, i))
+       clib_callback_enable_disable
+         (vlib_mains[i]->worker_thread_main_loop_callbacks,
+          vlib_mains[i]->worker_thread_main_loop_callback_tmp,
+          vlib_mains[i]->worker_thread_main_loop_callback_lock,
+          (void *) worker_thread_start_event, 1 /* enable */ );
     }
 }
 
@@ -397,23 +460,52 @@ scrape_and_clear_counters (perfmon_main_t * pm)
 }
 
 static void
-handle_timeout (perfmon_main_t * pm, f64 now)
+handle_timeout (vlib_main_t * vm, perfmon_main_t * pm, f64 now)
 {
   int i;
-  disable_events (pm);
+  int last_set, all;
+
+  last_set = clib_bitmap_last_set (pm->thread_bitmap);
+  all = (last_set == ~0);
+
+  if (all || clib_bitmap_get (pm->thread_bitmap, 0))
+    disable_events (pm);
 
   /* And also on worker threads */
   for (i = 1; i < vec_len (vlib_mains); i++)
     {
       if (vlib_mains[i] == 0)
        continue;
-      vlib_mains[i]->worker_thread_main_loop_callback = (void *)
-       worker_thread_stop_event;
+      if (all || clib_bitmap_get (pm->thread_bitmap, i))
+       clib_callback_enable_disable
+         (vlib_mains[i]->worker_thread_main_loop_callbacks,
+          vlib_mains[i]->worker_thread_main_loop_callback_tmp,
+          vlib_mains[i]->worker_thread_main_loop_callback_lock,
+          (void *) worker_thread_stop_event, 1 /* enable */ );
     }
 
-  /* Short delay to make sure workers have stopped collection */
+  /* Make sure workers have stopped collection */
   if (i > 1)
-    vlib_process_suspend (pm->vlib_main, 1e-3);
+    {
+      f64 deadman = vlib_time_now (vm) + 1.0;
+
+      for (i = 1; i < vec_len (vlib_mains); i++)
+       {
+         /* Has the worker actually stopped collecting data? */
+         while (clib_callback_is_set
+                (vlib_mains[i]->worker_thread_main_loop_callbacks,
+                 vlib_mains[i]->worker_thread_main_loop_callback_lock,
+                 read_current_perf_counters))
+           {
+             if (vlib_time_now (vm) > deadman)
+               {
+                 clib_warning ("Thread %d deadman timeout!", i);
+                 break;
+               }
+             vlib_process_suspend (pm->vlib_main, 1e-3);
+           }
+       }
+    }
   scrape_and_clear_counters (pm);
   pm->current_event += pm->n_active;
   if (pm->current_event >= vec_len (pm->single_events_to_collect))
@@ -422,15 +514,21 @@ handle_timeout (perfmon_main_t * pm, f64 now)
       pm->state = PERFMON_STATE_OFF;
       return;
     }
-  enable_current_events (pm);
+
+  if (all || clib_bitmap_get (pm->thread_bitmap, 0))
+    enable_current_events (pm);
 
   /* And also on worker threads */
   for (i = 1; i < vec_len (vlib_mains); i++)
     {
       if (vlib_mains[i] == 0)
        continue;
-      vlib_mains[i]->worker_thread_main_loop_callback = (void *)
-       worker_thread_start_event;
+      if (all || clib_bitmap_get (pm->thread_bitmap, i))
+       clib_callback_enable_disable
+         (vlib_mains[i]->worker_thread_main_loop_callbacks,
+          vlib_mains[i]->worker_thread_main_loop_callback_tmp,
+          vlib_mains[i]->worker_thread_main_loop_callback_lock,
+          worker_thread_start_event, 1 /* enable */ );
     }
 }
 
@@ -464,7 +562,7 @@ perfmon_periodic_process (vlib_main_t * vm,
 
          /* Handle timeout */
        case ~0:
-         handle_timeout (pm, now);
+         handle_timeout (vm, pm, now);
          break;
 
        default: