From: Florin Coras Date: Wed, 28 Jul 2021 05:48:05 +0000 (-0700) Subject: vcl: move vls pool lock to process local state X-Git-Tag: v22.02-rc0~160 X-Git-Url: https://gerrit.fd.io/r/gitweb?a=commitdiff_plain;h=2e2f9df2f157b60d30a4aa6256fefe793d186d4c;p=vpp.git vcl: move vls pool lock to process local state We only support one vls worker per process and therefore should not share lock between processes. Type: improvement Signed-off-by: Florin Coras Change-Id: I57bb536cf3bf04e8de031b07cb885f80b4fa03c9 --- diff --git a/src/vcl/vcl_locked.c b/src/vcl/vcl_locked.c index b006d89391b..6027fd19c15 100644 --- a/src/vcl/vcl_locked.c +++ b/src/vcl/vcl_locked.c @@ -118,6 +118,7 @@ typedef struct vls_local_ { int vls_wrk_index; /**< vls wrk index, 1 per process */ volatile int vls_mt_n_threads; /**< number of threads detected */ + clib_rwlock_t vls_pool_lock; /**< per process/wrk vls pool locks */ pthread_mutex_t vls_mt_mq_mlock; /**< vcl mq lock */ pthread_mutex_t vls_mt_spool_mlock; /**< vcl select or pool lock */ volatile u8 select_mp_check; /**< flag set if select checks done */ @@ -129,7 +130,6 @@ static vls_process_local_t *vlsl = &vls_local; typedef struct vls_main_ { vls_worker_t *workers; /**< pool of vls workers */ - clib_rwlock_t vls_pool_lock; /**< per process/wrk vls pool locks */ vls_shared_data_t *shared_data_pool; /**< inter proc pool of shared data */ clib_rwlock_t shared_data_lock; /**< shared data pool lock */ clib_spinlock_t worker_rpc_lock; /**< lock for inter-worker rpcs */ @@ -247,28 +247,28 @@ static inline void vls_mt_pool_rlock (void) { if (vlsl->vls_mt_n_threads > 1) - clib_rwlock_reader_lock (&vlsm->vls_pool_lock); + clib_rwlock_reader_lock (&vlsl->vls_pool_lock); } static inline void vls_mt_pool_runlock (void) { if (vlsl->vls_mt_n_threads > 1) - clib_rwlock_reader_unlock (&vlsm->vls_pool_lock); + clib_rwlock_reader_unlock (&vlsl->vls_pool_lock); } static inline void vls_mt_pool_wlock (void) { if (vlsl->vls_mt_n_threads > 1) - clib_rwlock_writer_lock (&vlsm->vls_pool_lock); + clib_rwlock_writer_lock (&vlsl->vls_pool_lock); } static inline void vls_mt_pool_wunlock (void) { if (vlsl->vls_mt_n_threads > 1) - clib_rwlock_writer_unlock (&vlsm->vls_pool_lock); + clib_rwlock_writer_unlock (&vlsl->vls_pool_lock); } typedef enum @@ -1769,6 +1769,7 @@ vls_app_fork_child_handler (void) vlsl->vls_mt_n_threads = 0; vlsl->vls_wrk_index = vcl_get_worker_index (); vlsl->select_mp_check = 0; + clib_rwlock_init (&vlsl->vls_pool_lock); vls_mt_locks_init (); parent_wrk = vcl_worker_get (parent_wrk_index); @@ -1940,7 +1941,6 @@ vls_app_create (char *app_name) vlsm = clib_mem_alloc (sizeof (vls_main_t)); clib_memset (vlsm, 0, sizeof (*vlsm)); - clib_rwlock_init (&vlsm->vls_pool_lock); clib_rwlock_init (&vlsm->shared_data_lock); clib_spinlock_init (&vlsm->worker_rpc_lock); pool_alloc (vlsm->workers, vcm->cfg.max_workers); @@ -1950,6 +1950,7 @@ vls_app_create (char *app_name) atexit (vls_app_exit); vls_worker_alloc (); vlsl->vls_wrk_index = vcl_get_worker_index (); + clib_rwlock_init (&vlsl->vls_pool_lock); vls_mt_locks_init (); vcm->wrk_rpc_fn = vls_rpc_handler; return VPPCOM_OK;