X-Git-Url: https://gerrit.fd.io/r/gitweb?p=vpp.git;a=blobdiff_plain;f=src%2Fvnet%2Fsession%2Fsession.h;h=0a72f97a2f82d73e1d45ed4243831a9e9f2cface;hp=92fd5fafe7a92d460630676fe53db4246c68ffed;hb=07063b8ea;hpb=f6c4313b6aa7746fe97afd398ce68c2efbef0600 diff --git a/src/vnet/session/session.h b/src/vnet/session/session.h index 92fd5fafe7a..0a72f97a2f8 100644 --- a/src/vnet/session/session.h +++ b/src/vnet/session/session.h @@ -15,13 +15,12 @@ #ifndef __included_session_h__ #define __included_session_h__ +#include #include #include #include -#include #include - -#define SESSION_PROXY_LISTENER_INDEX ((u8)~0 - 1) +#include #define foreach_session_input_error \ _(NO_SESSION, "No session drops") \ @@ -49,22 +48,30 @@ typedef struct session_tx_context_ session_t *s; transport_proto_vft_t *transport_vft; transport_connection_t *tc; - vlib_buffer_t *b; + transport_send_params_t sp; u32 max_dequeue; - u32 snd_space; u32 left_to_snd; - u32 tx_offset; u32 max_len_to_snd; u16 deq_per_first_buf; u16 deq_per_buf; - u16 snd_mss; u16 n_segs_per_evt; u8 n_bufs_per_seg; CLIB_CACHE_LINE_ALIGN_MARK (cacheline1); session_dgram_hdr_t hdr; } session_tx_context_t; -typedef struct session_manager_worker_ +typedef struct session_evt_elt +{ + clib_llist_anchor_t evt_list; + session_event_t evt; +} session_evt_elt_t; + +typedef struct session_ctrl_evt_data_ +{ + u8 data[SESSION_CTRL_MSG_MAX_SIZE]; +} session_evt_ctrl_data_t; + +typedef struct session_worker_ { CLIB_CACHE_LINE_ALIGN_MARK (cacheline0); @@ -74,17 +81,17 @@ typedef struct session_manager_worker_ /** vpp event message queue for worker */ svm_msg_q_t *vpp_event_queue; - /** Our approximation of a "complete" dispatch loop period */ - f64 dispatch_period; - /** vlib_time_now last time around the track */ - f64 last_vlib_time; + clib_time_type_t last_vlib_time; + + /** vlib_time_now rounded to us precision and as u64 */ + clib_us_time_t last_vlib_us_time; - /** Per-proto enqueue epoch counters */ - u64 current_enqueue_epoch[TRANSPORT_N_PROTO]; + /** Convenience pointer to this worker's vlib_main */ + vlib_main_t *vm; /** Per-proto vector of sessions to enqueue */ - u32 *session_to_enqueue[TRANSPORT_N_PROTO]; + u32 **session_to_enqueue; /** Context for session tx */ session_tx_context_t ctx; @@ -92,29 +99,39 @@ typedef struct session_manager_worker_ /** Vector of tx buffer free lists */ u32 *tx_buffers; - /** Vector of partially read events */ - session_event_t *free_event_vector; + /** Pool of session event list elements */ + session_evt_elt_t *event_elts; + + /** Pool of ctrl events data buffers */ + session_evt_ctrl_data_t *ctrl_evts_data; - /** Vector of active event vectors */ - session_event_t *pending_event_vector; + /** Head of control events list */ + clib_llist_index_t ctrl_head; - /** Vector of postponed disconnects */ - session_event_t *pending_disconnects; + /** Head of list of elements */ + clib_llist_index_t new_head; - /** Vector of postponed events */ - session_event_t *postponed_event_vector; + /** Head of list of pending events */ + clib_llist_index_t old_head; /** Peekers rw lock */ clib_rwlock_t peekers_rw_locks; - u32 last_tx_packets; + /** Vector of buffers to be sent */ + u32 *pending_tx_buffers; -} session_manager_worker_t; + /** Vector of nexts for the pending tx buffers */ + u16 *pending_tx_nexts; + +#if SESSION_DEBUG + /** last event poll time by thread */ + clib_time_type_t last_event_poll; +#endif +} session_worker_t; -typedef int (session_fifo_rx_fn) (vlib_main_t * vm, +typedef int (session_fifo_rx_fn) (session_worker_t * wrk, vlib_node_runtime_t * node, - session_manager_worker_t * wrk, - session_event_t * e, int *n_tx_pkts); + session_evt_elt_t * e, int *n_tx_packets); extern session_fifo_rx_fn session_tx_fifo_peek_and_snd; extern session_fifo_rx_fn session_tx_fifo_dequeue_and_snd; @@ -122,10 +139,10 @@ extern session_fifo_rx_fn session_tx_fifo_dequeue_internal; u8 session_node_lookup_fifo_event (svm_fifo_t * f, session_event_t * e); -typedef struct session_manager_main_ +typedef struct session_main_ { /** Worker contexts */ - session_manager_worker_t *wrk; + session_worker_t *wrk; /** Event queues memfd segment initialized only if so configured */ ssvm_private_t evt_qs_segment; @@ -141,12 +158,16 @@ typedef struct session_manager_main_ * Trade memory for speed, for now */ u32 *session_type_to_next; + transport_proto_t last_transport_proto_type; + /* * Config parameters */ /** Session manager is enabled */ u8 is_enabled; + /** Enable session manager at startup */ + u8 session_enable_asap; /** vpp fifo event queue configured length */ u32 configured_event_queue_length; @@ -174,68 +195,129 @@ typedef struct session_manager_main_ /** Preallocate session config parameter */ u32 preallocated_sessions; -#if SESSION_DEBUG - /** - * last event poll time by thread - * Debug only. Will cause false cache-line sharing as-is - */ - f64 *last_event_poll_by_thread; -#endif - -} session_manager_main_t; +} session_main_t; -extern session_manager_main_t session_manager_main; +extern session_main_t session_main; extern vlib_node_registration_t session_queue_node; extern vlib_node_registration_t session_queue_process_node; +extern vlib_node_registration_t session_queue_pre_input_node; #define SESSION_Q_PROCESS_FLUSH_FRAMES 1 #define SESSION_Q_PROCESS_STOP 2 -always_inline u8 -stream_session_is_valid (u32 si, u8 thread_index) +#define TRANSPORT_PROTO_INVALID (session_main.last_transport_proto_type + 1) +#define TRANSPORT_N_PROTOS (session_main.last_transport_proto_type + 1) + +static inline session_evt_elt_t * +session_evt_elt_alloc (session_worker_t * wrk) { - session_t *s; - s = pool_elt_at_index (session_manager_main.wrk[thread_index].sessions, si); - if (s->thread_index != thread_index || s->session_index != si - /* || s->server_rx_fifo->master_session_index != si - || s->server_tx_fifo->master_session_index != si - || s->server_rx_fifo->master_thread_index != thread_index - || s->server_tx_fifo->master_thread_index != thread_index */ ) - return 0; - return 1; + session_evt_elt_t *elt; + pool_get (wrk->event_elts, elt); + return elt; +} + +static inline void +session_evt_elt_free (session_worker_t * wrk, session_evt_elt_t * elt) +{ + pool_put (wrk->event_elts, elt); +} + +static inline void +session_evt_add_old (session_worker_t * wrk, session_evt_elt_t * elt) +{ + clib_llist_add_tail (wrk->event_elts, evt_list, elt, + pool_elt_at_index (wrk->event_elts, wrk->old_head)); +} + +static inline void +session_evt_add_head_old (session_worker_t * wrk, session_evt_elt_t * elt) +{ + clib_llist_add (wrk->event_elts, evt_list, elt, + pool_elt_at_index (wrk->event_elts, wrk->old_head)); +} + + +static inline u32 +session_evt_ctrl_data_alloc (session_worker_t * wrk) +{ + session_evt_ctrl_data_t *data; + pool_get (wrk->ctrl_evts_data, data); + return (data - wrk->ctrl_evts_data); +} + +static inline session_evt_elt_t * +session_evt_alloc_ctrl (session_worker_t * wrk) +{ + session_evt_elt_t *elt; + elt = session_evt_elt_alloc (wrk); + clib_llist_add_tail (wrk->event_elts, evt_list, elt, + pool_elt_at_index (wrk->event_elts, wrk->ctrl_head)); + return elt; +} + +static inline void * +session_evt_ctrl_data (session_worker_t * wrk, session_evt_elt_t * elt) +{ + return (void *) (pool_elt_at_index (wrk->ctrl_evts_data, + elt->evt.ctrl_data_index)); +} + +static inline void +session_evt_ctrl_data_free (session_worker_t * wrk, session_evt_elt_t * elt) +{ + ASSERT (elt->evt.event_type > SESSION_IO_EVT_BUILTIN_TX); + pool_put_index (wrk->ctrl_evts_data, elt->evt.ctrl_data_index); +} + +static inline session_evt_elt_t * +session_evt_alloc_new (session_worker_t * wrk) +{ + session_evt_elt_t *elt; + elt = session_evt_elt_alloc (wrk); + clib_llist_add_tail (wrk->event_elts, evt_list, elt, + pool_elt_at_index (wrk->event_elts, wrk->new_head)); + return elt; +} + +static inline session_evt_elt_t * +session_evt_alloc_old (session_worker_t * wrk) +{ + session_evt_elt_t *elt; + elt = session_evt_elt_alloc (wrk); + clib_llist_add_tail (wrk->event_elts, evt_list, elt, + pool_elt_at_index (wrk->event_elts, wrk->old_head)); + return elt; } session_t *session_alloc (u32 thread_index); void session_free (session_t * s); void session_free_w_fifos (session_t * s); +u8 session_is_valid (u32 si, u8 thread_index); always_inline session_t * session_get (u32 si, u32 thread_index) { - ASSERT (stream_session_is_valid (si, thread_index)); - return pool_elt_at_index (session_manager_main.wrk[thread_index].sessions, - si); + ASSERT (session_is_valid (si, thread_index)); + return pool_elt_at_index (session_main.wrk[thread_index].sessions, si); } always_inline session_t * session_get_if_valid (u64 si, u32 thread_index) { - if (thread_index >= vec_len (session_manager_main.wrk)) + if (thread_index >= vec_len (session_main.wrk)) return 0; - if (pool_is_free_index (session_manager_main.wrk[thread_index].sessions, - si)) + if (pool_is_free_index (session_main.wrk[thread_index].sessions, si)) return 0; - ASSERT (stream_session_is_valid (si, thread_index)); - return pool_elt_at_index (session_manager_main.wrk[thread_index].sessions, - si); + ASSERT (session_is_valid (si, thread_index)); + return pool_elt_at_index (session_main.wrk[thread_index].sessions, si); } always_inline session_t * session_get_from_handle (session_handle_t handle) { - session_manager_main_t *smm = &session_manager_main; + session_main_t *smm = &session_main; u32 session_index, thread_index; session_parse_handle (handle, &session_index, &thread_index); return pool_elt_at_index (smm->wrk[thread_index].sessions, session_index); @@ -249,18 +331,7 @@ session_get_from_handle_if_valid (session_handle_t handle) return session_get_if_valid (session_index, thread_index); } -always_inline u64 -session_segment_handle (session_t * s) -{ - svm_fifo_t *f; - - if (s->session_state == SESSION_STATE_LISTENING) - return SESSION_INVALID_HANDLE; - - f = s->rx_fifo; - return segment_manager_make_segment_handle (f->segment_manager, - f->segment_index); -} +u64 session_segment_handle (session_t * s); /** * Acquires a lock that blocks a session pool from expanding. @@ -275,7 +346,7 @@ session_segment_handle (session_t * s) always_inline void session_pool_add_peeker (u32 thread_index) { - session_manager_worker_t *wrk = &session_manager_main.wrk[thread_index]; + session_worker_t *wrk = &session_main.wrk[thread_index]; if (thread_index == vlib_get_thread_index ()) return; clib_rwlock_reader_lock (&wrk->peekers_rw_locks); @@ -284,7 +355,7 @@ session_pool_add_peeker (u32 thread_index) always_inline void session_pool_remove_peeker (u32 thread_index) { - session_manager_worker_t *wrk = &session_manager_main.wrk[thread_index]; + session_worker_t *wrk = &session_main.wrk[thread_index]; if (thread_index == vlib_get_thread_index ()) return; clib_rwlock_reader_unlock (&wrk->peekers_rw_locks); @@ -299,7 +370,7 @@ always_inline session_t * session_get_from_handle_safe (u64 handle) { u32 thread_index = session_thread_from_handle (handle); - session_manager_worker_t *wrk = &session_manager_main.wrk[thread_index]; + session_worker_t *wrk = &session_main.wrk[thread_index]; if (thread_index == vlib_get_thread_index ()) { @@ -317,7 +388,7 @@ session_get_from_handle_safe (u64 handle) always_inline u32 session_get_index (session_t * s) { - return (s - session_manager_main.wrk[s->thread_index].sessions); + return (s - session_main.wrk[s->thread_index].sessions); } always_inline session_t * @@ -333,7 +404,7 @@ session_clone_safe (u32 session_index, u32 thread_index) */ session_pool_add_peeker (thread_index); new_s = session_alloc (current_thread_index); - old_s = session_manager_main.wrk[thread_index].sessions + session_index; + old_s = session_main.wrk[thread_index].sessions + session_index; clib_memcpy_fast (new_s, old_s, sizeof (*new_s)); session_pool_remove_peeker (thread_index); new_s->thread_index = current_thread_index; @@ -345,20 +416,29 @@ int session_open (u32 app_index, session_endpoint_t * tep, u32 opaque); int session_listen (session_t * s, session_endpoint_cfg_t * sep); int session_stop_listen (session_t * s); void session_close (session_t * s); +void session_reset (session_t * s); void session_transport_close (session_t * s); +void session_transport_reset (session_t * s); void session_transport_cleanup (session_t * s); int session_send_io_evt_to_thread (svm_fifo_t * f, session_evt_type_t evt_type); +int session_enqueue_notify (session_t * s); int session_dequeue_notify (session_t * s); int session_send_io_evt_to_thread_custom (void *data, u32 thread_index, session_evt_type_t evt_type); void session_send_rpc_evt_to_thread (u32 thread_index, void *fp, void *rpc_args); +void session_send_rpc_evt_to_thread_force (u32 thread_index, void *fp, + void *rpc_args); +void session_add_self_custom_tx_evt (transport_connection_t * tc, + u8 has_prio); +void sesssion_reschedule_tx (transport_connection_t * tc); transport_connection_t *session_get_transport (session_t * s); +void session_get_endpoint (session_t * s, transport_endpoint_t * tep, + u8 is_lcl); - -u8 *format_stream_session (u8 * s, va_list * args); -uword unformat_stream_session (unformat_input_t * input, va_list * args); +u8 *format_session (u8 * s, va_list * args); +uword unformat_session (unformat_input_t * input, va_list * args); uword unformat_transport_connection (unformat_input_t * input, va_list * args); @@ -373,55 +453,70 @@ int session_enqueue_dgram_connection (session_t * s, session_dgram_hdr_t * hdr, vlib_buffer_t * b, u8 proto, u8 queue_event); -int stream_session_peek_bytes (transport_connection_t * tc, u8 * buffer, - u32 offset, u32 max_bytes); -u32 session_dequeue_drop (transport_connection_t * tc, u32 max_bytes); - int session_stream_connect_notify (transport_connection_t * tc, u8 is_fail); int session_dgram_connect_notify (transport_connection_t * tc, u32 old_thread_index, session_t ** new_session); -void stream_session_init_fifos_pointers (transport_connection_t * tc, - u32 rx_pointer, u32 tx_pointer); - int session_stream_accept_notify (transport_connection_t * tc); void session_transport_closing_notify (transport_connection_t * tc); void session_transport_delete_notify (transport_connection_t * tc); void session_transport_closed_notify (transport_connection_t * tc); void session_transport_reset_notify (transport_connection_t * tc); int session_stream_accept (transport_connection_t * tc, u32 listener_index, - u8 notify); -u32 session_tx_fifo_max_dequeue (transport_connection_t * tc); + u32 thread_index, u8 notify); +/** + * Initialize session layer for given transport proto and ip version + * + * Allocates per session type (transport proto + ip version) data structures + * and adds arc from session queue node to session type output node. + * + * @param transport_proto transport proto to be registered + * @param vft virtual function table for transport + * @param is_ip4 flag that indicates if transports uses ipv4 + * as underlying network layer + * @param output_node output node for transport + */ void session_register_transport (transport_proto_t transport_proto, const transport_proto_vft_t * vft, u8 is_ip4, u32 output_node); +transport_proto_t session_add_transport_proto (void); +int session_tx_fifo_peek_bytes (transport_connection_t * tc, u8 * buffer, + u32 offset, u32 max_bytes); +u32 session_tx_fifo_dequeue_drop (transport_connection_t * tc, u32 max_bytes); always_inline u32 transport_max_rx_enqueue (transport_connection_t * tc) { session_t *s = session_get (tc->s_index, tc->thread_index); - return svm_fifo_max_enqueue (s->rx_fifo); + return svm_fifo_max_enqueue_prod (s->rx_fifo); } always_inline u32 transport_max_tx_dequeue (transport_connection_t * tc) { session_t *s = session_get (tc->s_index, tc->thread_index); - return svm_fifo_max_dequeue (s->tx_fifo); + return svm_fifo_max_dequeue_cons (s->tx_fifo); +} + +always_inline u32 +transport_max_rx_dequeue (transport_connection_t * tc) +{ + session_t *s = session_get (tc->s_index, tc->thread_index); + return svm_fifo_max_dequeue (s->rx_fifo); } always_inline u32 transport_rx_fifo_size (transport_connection_t * tc) { session_t *s = session_get (tc->s_index, tc->thread_index); - return s->rx_fifo->nitems; + return svm_fifo_size (s->rx_fifo); } always_inline u32 transport_tx_fifo_size (transport_connection_t * tc) { session_t *s = session_get (tc->s_index, tc->thread_index); - return s->tx_fifo->nitems; + return svm_fifo_size (s->tx_fifo); } always_inline u8 @@ -431,16 +526,16 @@ transport_rx_fifo_has_ooo_data (transport_connection_t * tc) return svm_fifo_has_ooo_data (s->rx_fifo); } -always_inline f64 -transport_dispatch_period (u32 thread_index) +always_inline clib_time_type_t +transport_time_now (u32 thread_index) { - return session_manager_main.wrk[thread_index].dispatch_period; + return session_main.wrk[thread_index].last_vlib_time; } -always_inline f64 -transport_time_now (u32 thread_index) +always_inline clib_us_time_t +transport_us_time_now (u32 thread_index) { - return session_manager_main.wrk[thread_index].last_vlib_time; + return session_main.wrk[thread_index].last_vlib_us_time; } always_inline void @@ -459,7 +554,8 @@ transport_add_tx_event (transport_connection_t * tc) always_inline u64 listen_session_get_handle (session_t * s) { - ASSERT (s->session_state == SESSION_STATE_LISTENING); + ASSERT (s->session_state == SESSION_STATE_LISTENING || + session_get_transport_proto (s) == TRANSPORT_PROTO_QUIC); return session_handle (s); } @@ -500,51 +596,66 @@ listen_session_free (session_t * s) transport_connection_t *listen_session_get_transport (session_t * s); -int -listen_session_get_local_session_endpoint (session_t * listener, - session_endpoint_t * sep); - /* - * Session manager functions + * Session layer functions */ -always_inline session_manager_main_t * -vnet_get_session_manager_main () +always_inline session_main_t * +vnet_get_session_main () { - return &session_manager_main; + return &session_main; } -always_inline session_manager_worker_t * -session_manager_get_worker (u32 thread_index) +always_inline session_worker_t * +session_main_get_worker (u32 thread_index) { - return &session_manager_main.wrk[thread_index]; + return &session_main.wrk[thread_index]; +} + +static inline session_worker_t * +session_main_get_worker_if_valid (u32 thread_index) +{ + if (thread_index > vec_len (session_main.wrk)) + return 0; + return &session_main.wrk[thread_index]; } always_inline svm_msg_q_t * -session_manager_get_vpp_event_queue (u32 thread_index) +session_main_get_vpp_event_queue (u32 thread_index) { - return session_manager_main.wrk[thread_index].vpp_event_queue; + return session_main.wrk[thread_index].vpp_event_queue; } always_inline u8 -session_manager_is_enabled () +session_main_is_enabled () { - return session_manager_main.is_enabled == 1; + return session_main.is_enabled == 1; } #define session_cli_return_if_not_enabled() \ do { \ - if (!session_manager_main.is_enabled) \ - return clib_error_return(0, "session layer is not enabled"); \ + if (!session_main.is_enabled) \ + return clib_error_return (0, "session layer is not enabled"); \ } while (0) -int session_manager_flush_enqueue_events (u8 proto, u32 thread_index); -int session_manager_flush_all_enqueue_events (u8 transport_proto); +int session_main_flush_enqueue_events (u8 proto, u32 thread_index); +int session_main_flush_all_enqueue_events (u8 transport_proto); void session_flush_frames_main_thread (vlib_main_t * vm); -ssvm_private_t *session_manager_get_evt_q_segment (void); + +always_inline void +session_add_pending_tx_buffer (session_type_t st, u32 thread_index, u32 bi) +{ + session_worker_t *wrk = session_main_get_worker (thread_index); + vec_add1 (wrk->pending_tx_buffers, bi); + vec_add1 (wrk->pending_tx_nexts, session_main.session_type_to_next[st]); +} + +ssvm_private_t *session_main_get_evt_q_segment (void); void session_node_enable_disable (u8 is_en); clib_error_t *vnet_session_enable_disable (vlib_main_t * vm, u8 is_en); +session_t *session_alloc_for_connection (transport_connection_t * tc); + #endif /* __included_session_h__ */ /*