vcl: wait for segments with segment handle
[vpp.git] / src / vcl / vcl_private.c
index 48abbaf..76ae0e7 100644 (file)
 
 #include <vcl/vcl_private.h>
 
+static pthread_key_t vcl_worker_stop_key;
+
+static const char *
+vppcom_app_state_str (app_state_t state)
+{
+  char *st;
+
+  switch (state)
+    {
+    case STATE_APP_START:
+      st = "STATE_APP_START";
+      break;
+
+    case STATE_APP_CONN_VPP:
+      st = "STATE_APP_CONN_VPP";
+      break;
+
+    case STATE_APP_ENABLED:
+      st = "STATE_APP_ENABLED";
+      break;
+
+    case STATE_APP_ATTACHED:
+      st = "STATE_APP_ATTACHED";
+      break;
+
+    default:
+      st = "UNKNOWN_APP_STATE";
+      break;
+    }
+
+  return st;
+}
+
+int
+vcl_wait_for_app_state_change (app_state_t app_state)
+{
+  vcl_worker_t *wrk = vcl_worker_get_current ();
+  f64 timeout = clib_time_now (&wrk->clib_time) + vcm->cfg.app_timeout;
+
+  while (clib_time_now (&wrk->clib_time) < timeout)
+    {
+      if (vcm->app_state == app_state)
+       return VPPCOM_OK;
+      if (vcm->app_state == STATE_APP_FAILED)
+       return VPPCOM_ECONNABORTED;
+    }
+  VDBG (0, "VCL<%d>: timeout waiting for state %s (%d)", getpid (),
+       vppcom_app_state_str (app_state), app_state);
+  vcl_evt (VCL_EVT_SESSION_TIMEOUT, vcm, app_state);
+
+  return VPPCOM_ETIMEDOUT;
+}
+
 vcl_cut_through_registration_t *
-vcl_ct_registration_lock_and_alloc (void)
+vcl_ct_registration_lock_and_alloc (vcl_worker_t * wrk)
 {
   vcl_cut_through_registration_t *cr;
-  pool_get (vcm->cut_through_registrations, cr);
-  clib_spinlock_lock (&vcm->ct_registration_lock);
+  clib_spinlock_lock (&wrk->ct_registration_lock);
+  pool_get (wrk->cut_through_registrations, cr);
   memset (cr, 0, sizeof (*cr));
   cr->epoll_evt_conn_index = -1;
   return cr;
 }
 
 u32
-vcl_ct_registration_index (vcl_cut_through_registration_t * ctr)
+vcl_ct_registration_index (vcl_worker_t * wrk,
+                          vcl_cut_through_registration_t * ctr)
+{
+  return (ctr - wrk->cut_through_registrations);
+}
+
+void
+vcl_ct_registration_lock (vcl_worker_t * wrk)
 {
-  return (ctr - vcm->cut_through_registrations);
+  clib_spinlock_lock (&wrk->ct_registration_lock);
 }
 
 void
-vcl_ct_registration_unlock (void)
+vcl_ct_registration_unlock (vcl_worker_t * wrk)
 {
-  clib_spinlock_unlock (&vcm->ct_registration_lock);
+  clib_spinlock_unlock (&wrk->ct_registration_lock);
 }
 
 vcl_cut_through_registration_t *
-vcl_ct_registration_get (u32 ctr_index)
+vcl_ct_registration_get (vcl_worker_t * wrk, u32 ctr_index)
 {
-  if (pool_is_free_index (vcm->cut_through_registrations, ctr_index))
+  if (pool_is_free_index (wrk->cut_through_registrations, ctr_index))
     return 0;
-  return pool_elt_at_index (vcm->cut_through_registrations, ctr_index);
+  return pool_elt_at_index (wrk->cut_through_registrations, ctr_index);
 }
 
 vcl_cut_through_registration_t *
-vcl_ct_registration_lock_and_lookup (uword mq_addr)
+vcl_ct_registration_lock_and_lookup (vcl_worker_t * wrk, uword mq_addr)
 {
   uword *p;
-  clib_spinlock_lock (&vcm->ct_registration_lock);
-  p = hash_get (vcm->ct_registration_by_mq, mq_addr);
+  clib_spinlock_lock (&wrk->ct_registration_lock);
+  p = hash_get (wrk->ct_registration_by_mq, mq_addr);
   if (!p)
     return 0;
-  return vcl_ct_registration_get (p[0]);
+  return vcl_ct_registration_get (wrk, p[0]);
 }
 
 void
-vcl_ct_registration_lookup_add (uword mq_addr, u32 ctr_index)
+vcl_ct_registration_lookup_add (vcl_worker_t * wrk, uword mq_addr,
+                               u32 ctr_index)
 {
-  hash_set (vcm->ct_registration_by_mq, mq_addr, ctr_index);
+  hash_set (wrk->ct_registration_by_mq, mq_addr, ctr_index);
 }
 
 void
-vcl_ct_registration_lookup_del (uword mq_addr)
+vcl_ct_registration_lookup_del (vcl_worker_t * wrk, uword mq_addr)
 {
-  hash_unset (vcm->ct_registration_by_mq, mq_addr);
+  hash_unset (wrk->ct_registration_by_mq, mq_addr);
 }
 
 void
-vcl_ct_registration_del (vcl_cut_through_registration_t * ctr)
+vcl_ct_registration_del (vcl_worker_t * wrk,
+                        vcl_cut_through_registration_t * ctr)
 {
-  pool_put (vcm->cut_through_registrations, ctr);
+  pool_put (wrk->cut_through_registrations, ctr);
 }
 
 vcl_mq_evt_conn_t *
-vcl_mq_evt_conn_alloc (void)
+vcl_mq_evt_conn_alloc (vcl_worker_t * wrk)
 {
   vcl_mq_evt_conn_t *mqc;
-  pool_get (vcm->mq_evt_conns, mqc);
+  pool_get (wrk->mq_evt_conns, mqc);
   memset (mqc, 0, sizeof (*mqc));
   return mqc;
 }
 
 u32
-vcl_mq_evt_conn_index (vcl_mq_evt_conn_t * mqc)
+vcl_mq_evt_conn_index (vcl_worker_t * wrk, vcl_mq_evt_conn_t * mqc)
 {
-  return (mqc - vcm->mq_evt_conns);
+  return (mqc - wrk->mq_evt_conns);
 }
 
 vcl_mq_evt_conn_t *
-vcl_mq_evt_conn_get (u32 mq_conn_idx)
+vcl_mq_evt_conn_get (vcl_worker_t * wrk, u32 mq_conn_idx)
 {
-  return pool_elt_at_index (vcm->mq_evt_conns, mq_conn_idx);
+  return pool_elt_at_index (wrk->mq_evt_conns, mq_conn_idx);
 }
 
 int
-vcl_mq_epoll_add_evfd (svm_msg_q_t * mq)
+vcl_mq_epoll_add_evfd (vcl_worker_t * wrk, svm_msg_q_t * mq)
 {
   struct epoll_event e = { 0 };
   vcl_mq_evt_conn_t *mqc;
@@ -106,17 +168,17 @@ vcl_mq_epoll_add_evfd (svm_msg_q_t * mq)
 
   mq_fd = svm_msg_q_get_consumer_eventfd (mq);
 
-  if (vcm->mqs_epfd < 0 || mq_fd == -1)
+  if (wrk->mqs_epfd < 0 || mq_fd == -1)
     return -1;
 
-  mqc = vcl_mq_evt_conn_alloc ();
-  mqc_index = vcl_mq_evt_conn_index (mqc);
+  mqc = vcl_mq_evt_conn_alloc (wrk);
+  mqc_index = vcl_mq_evt_conn_index (wrk, mqc);
   mqc->mq_fd = mq_fd;
   mqc->mq = mq;
 
   e.events = EPOLLIN;
   e.data.u32 = mqc_index;
-  if (epoll_ctl (vcm->mqs_epfd, EPOLL_CTL_ADD, mq_fd, &e) < 0)
+  if (epoll_ctl (wrk->mqs_epfd, EPOLL_CTL_ADD, mq_fd, &e) < 0)
     {
       clib_warning ("failed to add mq eventfd to mq epoll fd");
       return -1;
@@ -126,15 +188,15 @@ vcl_mq_epoll_add_evfd (svm_msg_q_t * mq)
 }
 
 int
-vcl_mq_epoll_del_evfd (u32 mqc_index)
+vcl_mq_epoll_del_evfd (vcl_worker_t * wrk, u32 mqc_index)
 {
   vcl_mq_evt_conn_t *mqc;
 
-  if (vcm->mqs_epfd || mqc_index == ~0)
+  if (wrk->mqs_epfd || mqc_index == ~0)
     return -1;
 
-  mqc = vcl_mq_evt_conn_get (mqc_index);
-  if (epoll_ctl (vcm->mqs_epfd, EPOLL_CTL_DEL, mqc->mq_fd, 0) < 0)
+  mqc = vcl_mq_evt_conn_get (wrk, mqc_index);
+  if (epoll_ctl (wrk->mqs_epfd, EPOLL_CTL_DEL, mqc->mq_fd, 0) < 0)
     {
       clib_warning ("failed to del mq eventfd to mq epoll fd");
       return -1;
@@ -142,6 +204,274 @@ vcl_mq_epoll_del_evfd (u32 mqc_index)
   return 0;
 }
 
+static vcl_worker_t *
+vcl_worker_alloc (void)
+{
+  vcl_worker_t *wrk;
+  pool_get (vcm->workers, wrk);
+  memset (wrk, 0, sizeof (*wrk));
+  wrk->wrk_index = wrk - vcm->workers;
+  return wrk;
+}
+
+static void
+vcl_worker_free (vcl_worker_t * wrk)
+{
+  pool_put (vcm->workers, wrk);
+}
+
+void
+vcl_worker_cleanup (void)
+{
+  vcl_worker_t *wrk = vcl_worker_get_current ();
+
+  clib_spinlock_lock (&vcm->workers_lock);
+  vcl_send_app_worker_add_del (0 /* is_add */ );
+  close (wrk->mqs_epfd);
+  hash_free (wrk->session_index_by_vpp_handles);
+  hash_free (wrk->ct_registration_by_mq);
+  clib_spinlock_free (&wrk->ct_registration_lock);
+  vec_free (wrk->mq_events);
+  vec_free (wrk->mq_msg_vector);
+  vcl_set_worker_index (~0);
+  vcl_worker_free (wrk);
+  clib_spinlock_unlock (&vcm->workers_lock);
+  VDBG (0, "cleaned up worker %u", wrk->wrk_index);
+}
+
+static void
+vcl_worker_cleanup_cb (void *arg)
+{
+  vcl_worker_cleanup ();
+}
+
+vcl_worker_t *
+vcl_worker_alloc_and_init ()
+{
+  vcl_worker_t *wrk;
+
+  /* This was initialized already */
+  if (vcl_get_worker_index () != ~0)
+    return 0;
+
+  if (pool_elts (vcm->workers) == vcm->cfg.max_workers)
+    {
+      VDBG (0, "max-workers %u limit reached", vcm->cfg.max_workers);
+      return 0;
+    }
+
+  clib_spinlock_lock (&vcm->workers_lock);
+  wrk = vcl_worker_alloc ();
+  vcl_set_worker_index (wrk->wrk_index);
+  wrk->thread_id = pthread_self ();
+  wrk->current_pid = getpid ();
+
+  wrk->mqs_epfd = -1;
+  if (vcm->cfg.use_mq_eventfd)
+    {
+      wrk->mqs_epfd = epoll_create (1);
+      if (wrk->mqs_epfd < 0)
+       {
+         clib_unix_warning ("epoll_create() returned");
+         goto done;
+       }
+    }
+
+  wrk->session_index_by_vpp_handles = hash_create (0, sizeof (uword));
+  wrk->ct_registration_by_mq = hash_create (0, sizeof (uword));
+  clib_spinlock_init (&wrk->ct_registration_lock);
+  clib_time_init (&wrk->clib_time);
+  vec_validate (wrk->mq_events, 64);
+  vec_validate (wrk->mq_msg_vector, 128);
+  vec_reset_length (wrk->mq_msg_vector);
+  vec_validate (wrk->unhandled_evts_vector, 128);
+  vec_reset_length (wrk->unhandled_evts_vector);
+  clib_spinlock_unlock (&vcm->workers_lock);
+
+done:
+  return wrk;
+}
+
+int
+vcl_worker_register_with_vpp (void)
+{
+  vcl_worker_t *wrk = vcl_worker_get_current ();
+
+  clib_spinlock_lock (&vcm->workers_lock);
+
+  vcm->app_state = STATE_APP_ADDING_WORKER;
+  vcl_send_app_worker_add_del (1 /* is_add */ );
+  if (vcl_wait_for_app_state_change (STATE_APP_READY))
+    {
+      clib_warning ("failed to add worker to vpp");
+      return -1;
+    }
+
+  if (pthread_key_create (&vcl_worker_stop_key, vcl_worker_cleanup_cb))
+    clib_warning ("failed to add pthread cleanup function");
+  if (pthread_setspecific (vcl_worker_stop_key, &wrk->thread_id))
+    clib_warning ("failed to setup key value");
+
+  clib_spinlock_unlock (&vcm->workers_lock);
+
+  VDBG (0, "added worker %u", wrk->wrk_index);
+  return 0;
+}
+
+int
+vcl_worker_set_bapi (void)
+{
+  vcl_worker_t *wrk = vcl_worker_get_current ();
+  int i;
+
+  /* Find the first worker with the same pid */
+  for (i = 0; i < vec_len (vcm->workers); i++)
+    {
+      if (i == wrk->wrk_index)
+       continue;
+      if (vcm->workers[i].current_pid == wrk->current_pid)
+       {
+         wrk->vl_input_queue = vcm->workers[i].vl_input_queue;
+         wrk->my_client_index = vcm->workers[i].my_client_index;
+         return 0;
+       }
+    }
+  return -1;
+}
+
+vcl_shared_session_t *
+vcl_shared_session_alloc (void)
+{
+  vcl_shared_session_t *ss;
+  pool_get (vcm->shared_sessions, ss);
+  memset (ss, 0, sizeof (*ss));
+  ss->ss_index = ss - vcm->shared_sessions;
+  return ss;
+}
+
+vcl_shared_session_t *
+vcl_shared_session_get (u32 ss_index)
+{
+  if (pool_is_free_index (vcm->shared_sessions, ss_index))
+    return 0;
+  return pool_elt_at_index (vcm->shared_sessions, ss_index);
+}
+
+void
+vcl_shared_session_free (vcl_shared_session_t * ss)
+{
+  pool_put (vcm->shared_sessions, ss);
+}
+
+void
+vcl_worker_share_session (vcl_worker_t * parent, vcl_worker_t * wrk,
+                         vcl_session_t * new_s)
+{
+  vcl_shared_session_t *ss;
+  vcl_session_t *s;
+
+  s = vcl_session_get (parent, new_s->session_index);
+  if (s->shared_index == ~0)
+    {
+      ss = vcl_shared_session_alloc ();
+      vec_add1 (ss->workers, parent->wrk_index);
+      s->shared_index = ss->ss_index;
+    }
+  else
+    {
+      ss = vcl_shared_session_get (s->shared_index);
+    }
+  new_s->shared_index = ss->ss_index;
+  vec_add1 (ss->workers, wrk->wrk_index);
+}
+
+int
+vcl_worker_unshare_session (vcl_worker_t * wrk, vcl_session_t * s)
+{
+  vcl_shared_session_t *ss;
+  int i;
+
+  ss = vcl_shared_session_get (s->shared_index);
+  for (i = 0; i < vec_len (ss->workers); i++)
+    {
+      if (ss->workers[i] == wrk->wrk_index)
+       {
+         vec_del1 (ss->workers, i);
+         break;
+       }
+    }
+
+  if (vec_len (ss->workers) == 0)
+    {
+      vcl_shared_session_free (ss);
+      return 1;
+    }
+
+  return 0;
+}
+
+void
+vcl_worker_share_sessions (u32 parent_wrk_index)
+{
+  vcl_worker_t *parent_wrk, *wrk;
+  vcl_session_t *new_s;
+
+  parent_wrk = vcl_worker_get (parent_wrk_index);
+  if (!parent_wrk->sessions)
+    return;
+
+  wrk = vcl_worker_get_current ();
+  wrk->sessions = pool_dup (parent_wrk->sessions);
+  wrk->session_index_by_vpp_handles =
+    hash_dup (parent_wrk->session_index_by_vpp_handles);
+
+  /* *INDENT-OFF* */
+  pool_foreach (new_s, wrk->sessions, ({
+    vcl_worker_share_session (parent_wrk, wrk, new_s);
+  }));
+  /* *INDENT-ON* */
+}
+
+int
+vcl_session_get_refcnt (vcl_session_t * s)
+{
+  vcl_shared_session_t *ss;
+  ss = vcl_shared_session_get (s->shared_index);
+  if (ss)
+    return vec_len (ss->workers);
+  return 0;
+}
+
+void
+vcl_segment_table_add (u64 segment_handle, u32 svm_segment_index)
+{
+  clib_rwlock_writer_lock (&vcm->segment_table_lock);
+  hash_set (vcm->segment_table, segment_handle, svm_segment_index);
+  clib_rwlock_writer_unlock (&vcm->segment_table_lock);
+}
+
+u32
+vcl_segment_table_lookup (u64 segment_handle)
+{
+  uword *seg_indexp;
+
+  clib_rwlock_reader_lock (&vcm->segment_table_lock);
+  seg_indexp = hash_get (vcm->segment_table, segment_handle);
+  clib_rwlock_reader_unlock (&vcm->segment_table_lock);
+
+  if (!seg_indexp)
+    return VCL_INVALID_SEGMENT_INDEX;
+  return ((u32) * seg_indexp);
+}
+
+void
+vcl_segment_table_del (u64 segment_handle)
+{
+  clib_rwlock_writer_lock (&vcm->segment_table_lock);
+  hash_unset (vcm->segment_table, segment_handle);
+  clib_rwlock_writer_unlock (&vcm->segment_table_lock);
+}
+
 /*
  * fd.io coding-style-patch-verification: ON
  *