X-Git-Url: https://gerrit.fd.io/r/gitweb?a=blobdiff_plain;f=src%2Fvcl%2Fvcl_private.c;h=300b82c4f21117b3691fc0175be44668b8e90614;hb=7a2abce4c9fc2191b3d3989c84516391036b6882;hp=6c364e376f7ed9598f04e6b05ceb31b5622d7fe5;hpb=288eaab5964b9211350acad8d742fae4789577fe;p=vpp.git diff --git a/src/vcl/vcl_private.c b/src/vcl/vcl_private.c index 6c364e376f7..300b82c4f21 100644 --- a/src/vcl/vcl_private.c +++ b/src/vcl/vcl_private.c @@ -1,5 +1,5 @@ /* - * Copyright (c) 2018 Cisco and/or its affiliates. + * Copyright (c) 2018-2019 Cisco and/or its affiliates. * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this * You may obtain a copy of the License at: @@ -61,82 +61,13 @@ vcl_wait_for_app_state_change (app_state_t app_state) if (vcm->app_state == STATE_APP_FAILED) return VPPCOM_ECONNABORTED; } - VDBG (0, "VCL<%d>: timeout waiting for state %s (%d)", getpid (), + VDBG (0, "timeout waiting for state %s (%d)", vppcom_app_state_str (app_state), app_state); vcl_evt (VCL_EVT_SESSION_TIMEOUT, vcm, app_state); return VPPCOM_ETIMEDOUT; } -vcl_cut_through_registration_t * -vcl_ct_registration_lock_and_alloc (vcl_worker_t * wrk) -{ - vcl_cut_through_registration_t *cr; - clib_spinlock_lock (&wrk->ct_registration_lock); - pool_get (wrk->cut_through_registrations, cr); - memset (cr, 0, sizeof (*cr)); - cr->epoll_evt_conn_index = -1; - return cr; -} - -u32 -vcl_ct_registration_index (vcl_worker_t * wrk, - vcl_cut_through_registration_t * ctr) -{ - return (ctr - wrk->cut_through_registrations); -} - -void -vcl_ct_registration_lock (vcl_worker_t * wrk) -{ - clib_spinlock_lock (&wrk->ct_registration_lock); -} - -void -vcl_ct_registration_unlock (vcl_worker_t * wrk) -{ - clib_spinlock_unlock (&wrk->ct_registration_lock); -} - -vcl_cut_through_registration_t * -vcl_ct_registration_get (vcl_worker_t * wrk, u32 ctr_index) -{ - if (pool_is_free_index (wrk->cut_through_registrations, ctr_index)) - return 0; - return pool_elt_at_index (wrk->cut_through_registrations, ctr_index); -} - -vcl_cut_through_registration_t * -vcl_ct_registration_lock_and_lookup (vcl_worker_t * wrk, uword mq_addr) -{ - uword *p; - clib_spinlock_lock (&wrk->ct_registration_lock); - p = hash_get (wrk->ct_registration_by_mq, mq_addr); - if (!p) - return 0; - return vcl_ct_registration_get (wrk, p[0]); -} - -void -vcl_ct_registration_lookup_add (vcl_worker_t * wrk, uword mq_addr, - u32 ctr_index) -{ - hash_set (wrk->ct_registration_by_mq, mq_addr, ctr_index); -} - -void -vcl_ct_registration_lookup_del (vcl_worker_t * wrk, uword mq_addr) -{ - hash_unset (wrk->ct_registration_by_mq, mq_addr); -} - -void -vcl_ct_registration_del (vcl_worker_t * wrk, - vcl_cut_through_registration_t * ctr) -{ - pool_put (wrk->cut_through_registrations, ctr); -} - vcl_mq_evt_conn_t * vcl_mq_evt_conn_alloc (vcl_worker_t * wrk) { @@ -180,7 +111,7 @@ vcl_mq_epoll_add_evfd (vcl_worker_t * wrk, svm_msg_q_t * mq) e.data.u32 = mqc_index; if (epoll_ctl (wrk->mqs_epfd, EPOLL_CTL_ADD, mq_fd, &e) < 0) { - clib_warning ("failed to add mq eventfd to mq epoll fd"); + VDBG (0, "failed to add mq eventfd to mq epoll fd"); return -1; } @@ -198,7 +129,7 @@ vcl_mq_epoll_del_evfd (vcl_worker_t * wrk, u32 mqc_index) mqc = vcl_mq_evt_conn_get (wrk, mqc_index); if (epoll_ctl (wrk->mqs_epfd, EPOLL_CTL_DEL, mqc->mq_fd, 0) < 0) { - clib_warning ("failed to del mq eventfd to mq epoll fd"); + VDBG (0, "failed to del mq eventfd to mq epoll fd"); return -1; } return 0; @@ -227,16 +158,22 @@ vcl_worker_cleanup (vcl_worker_t * wrk, u8 notify_vpp) clib_spinlock_lock (&vcm->workers_lock); if (notify_vpp) { + /* Notify vpp that the worker is going away */ if (wrk->wrk_index == vcl_get_worker_index ()) vcl_send_app_worker_add_del (0 /* is_add */ ); else vcl_send_child_worker_del (wrk); + + /* Disconnect the binary api */ + if (vec_len (vcm->workers) == 1) + vppcom_disconnect_from_vpp (); + else + vl_client_send_disconnect (1 /* vpp should cleanup */ ); } + if (wrk->mqs_epfd > 0) close (wrk->mqs_epfd); hash_free (wrk->session_index_by_vpp_handles); - hash_free (wrk->ct_registration_by_mq); - clib_spinlock_free (&wrk->ct_registration_lock); vec_free (wrk->mq_events); vec_free (wrk->mq_msg_vector); vcl_worker_free (wrk); @@ -262,6 +199,9 @@ vcl_worker_alloc_and_init () if (vcl_get_worker_index () != ~0) return 0; + /* Use separate heap map entry for worker */ + clib_mem_set_thread_index (); + if (pool_elts (vcm->workers) == vcm->cfg.max_workers) { VDBG (0, "max-workers %u limit reached", vcm->cfg.max_workers); @@ -286,8 +226,6 @@ vcl_worker_alloc_and_init () } wrk->session_index_by_vpp_handles = hash_create (0, sizeof (uword)); - wrk->ct_registration_by_mq = hash_create (0, sizeof (uword)); - clib_spinlock_init (&wrk->ct_registration_lock); clib_time_init (&wrk->clib_time); vec_validate (wrk->mq_events, 64); vec_validate (wrk->mq_msg_vector, 128); @@ -311,13 +249,13 @@ vcl_worker_register_with_vpp (void) vcl_send_app_worker_add_del (1 /* is_add */ ); if (vcl_wait_for_app_state_change (STATE_APP_READY)) { - clib_warning ("failed to add worker to vpp"); + VDBG (0, "failed to add worker to vpp"); return -1; } if (pthread_key_create (&vcl_worker_stop_key, vcl_worker_cleanup_cb)) - clib_warning ("failed to add pthread cleanup function"); + VDBG (0, "failed to add pthread cleanup function"); if (pthread_setspecific (vcl_worker_stop_key, &wrk->thread_id)) - clib_warning ("failed to setup key value"); + VDBG (0, "failed to setup key value"); clib_spinlock_unlock (&vcm->workers_lock); @@ -346,41 +284,17 @@ vcl_worker_set_bapi (void) return -1; } -void -vcl_segment_table_add (u64 segment_handle, u32 svm_segment_index) -{ - clib_rwlock_writer_lock (&vcm->segment_table_lock); - hash_set (vcm->segment_table, segment_handle, svm_segment_index); - clib_rwlock_writer_unlock (&vcm->segment_table_lock); -} - -u32 -vcl_segment_table_lookup (u64 segment_handle) +svm_msg_q_t * +vcl_worker_ctrl_mq (vcl_worker_t * wrk) { - uword *seg_indexp; - - clib_rwlock_reader_lock (&vcm->segment_table_lock); - seg_indexp = hash_get (vcm->segment_table, segment_handle); - clib_rwlock_reader_unlock (&vcm->segment_table_lock); - - if (!seg_indexp) - return VCL_INVALID_SEGMENT_INDEX; - return ((u32) * seg_indexp); -} - -void -vcl_segment_table_del (u64 segment_handle) -{ - clib_rwlock_writer_lock (&vcm->segment_table_lock); - hash_unset (vcm->segment_table, segment_handle); - clib_rwlock_writer_unlock (&vcm->segment_table_lock); + return wrk->ctrl_mq; } void vcl_cleanup_bapi (void) { socket_client_main_t *scm = &socket_client_main; - api_main_t *am = &api_main; + api_main_t *am = vlibapi_get_main (); am->my_client_index = ~0; am->my_registration = 0; @@ -394,6 +308,8 @@ vcl_cleanup_bapi (void) int vcl_session_read_ready (vcl_session_t * session) { + u32 max_deq; + /* Assumes caller has acquired spinlock: vcm->sessions_lockp */ if (PREDICT_FALSE (session->is_vep)) { @@ -418,7 +334,26 @@ vcl_session_read_ready (vcl_session_t * session) if (session->session_state & STATE_LISTEN) return clib_fifo_elts (session->accept_evts_fifo); - return svm_fifo_max_dequeue (session->rx_fifo); + if (vcl_session_is_ct (session)) + return svm_fifo_max_dequeue_cons (session->ct_rx_fifo); + + max_deq = svm_fifo_max_dequeue_cons (session->rx_fifo); + + if (session->is_dgram) + { + session_dgram_pre_hdr_t ph; + + if (max_deq <= SESSION_CONN_HDR_LEN) + return 0; + if (svm_fifo_peek (session->rx_fifo, 0, sizeof (ph), (u8 *) & ph) < 0) + return 0; + if (ph.data_length + SESSION_CONN_HDR_LEN > max_deq) + return 0; + + return ph.data_length; + } + + return max_deq; } int @@ -435,7 +370,7 @@ vcl_session_write_ready (vcl_session_t * session) if (PREDICT_FALSE (session->session_state & STATE_LISTEN)) { if (session->tx_fifo) - return svm_fifo_max_enqueue (session->tx_fifo); + return svm_fifo_max_enqueue_prod (session->tx_fifo); else return VPPCOM_EBADFD; } @@ -452,9 +387,87 @@ vcl_session_write_ready (vcl_session_t * session) return rv; } - return svm_fifo_max_enqueue (session->tx_fifo); + if (vcl_session_is_ct (session)) + return svm_fifo_max_enqueue_prod (session->ct_tx_fifo); + + if (session->is_dgram) + { + u32 max_enq = svm_fifo_max_enqueue_prod (session->tx_fifo); + + if (max_enq <= sizeof (session_dgram_hdr_t)) + return 0; + return max_enq - sizeof (session_dgram_hdr_t); + } + + return svm_fifo_max_enqueue_prod (session->tx_fifo); } +int +vcl_segment_attach (u64 segment_handle, char *name, ssvm_segment_type_t type, + int fd) +{ + fifo_segment_create_args_t _a, *a = &_a; + int rv; + + memset (a, 0, sizeof (*a)); + a->segment_name = name; + a->segment_type = type; + + if (type == SSVM_SEGMENT_MEMFD) + a->memfd_fd = fd; + + clib_rwlock_writer_lock (&vcm->segment_table_lock); + + if ((rv = fifo_segment_attach (&vcm->segment_main, a))) + { + clib_warning ("svm_fifo_segment_attach ('%s') failed", name); + return rv; + } + hash_set (vcm->segment_table, segment_handle, a->new_segment_indices[0]); + + clib_rwlock_writer_unlock (&vcm->segment_table_lock); + + vec_reset_length (a->new_segment_indices); + return 0; +} + +u32 +vcl_segment_table_lookup (u64 segment_handle) +{ + uword *seg_indexp; + + clib_rwlock_reader_lock (&vcm->segment_table_lock); + seg_indexp = hash_get (vcm->segment_table, segment_handle); + clib_rwlock_reader_unlock (&vcm->segment_table_lock); + + if (!seg_indexp) + return VCL_INVALID_SEGMENT_INDEX; + return ((u32) * seg_indexp); +} + +void +vcl_segment_detach (u64 segment_handle) +{ + fifo_segment_main_t *sm = &vcm->segment_main; + fifo_segment_t *segment; + u32 segment_index; + + segment_index = vcl_segment_table_lookup (segment_handle); + if (segment_index == (u32) ~ 0) + return; + + clib_rwlock_writer_lock (&vcm->segment_table_lock); + + segment = fifo_segment_get_segment (sm, segment_index); + fifo_segment_delete (sm, segment); + hash_unset (vcm->segment_table, segment_handle); + + clib_rwlock_writer_unlock (&vcm->segment_table_lock); + + VDBG (0, "detached segment %u handle %u", segment_index, segment_handle); +} + + /* * fd.io coding-style-patch-verification: ON *