X-Git-Url: https://gerrit.fd.io/r/gitweb?p=vpp.git;a=blobdiff_plain;f=src%2Fvnet%2Fsession%2Fsession.h;h=d5f040edc594fa84a8fad637ab7432b2b78eed23;hp=5fa4225c4ed2332fe3fbf91898a3be99d0816bde;hb=fa76a76;hpb=f03a59ab008908f98fd7d1b187a8c0fb78b01add diff --git a/src/vnet/session/session.h b/src/vnet/session/session.h index 5fa4225c4ed..d5f040edc59 100644 --- a/src/vnet/session/session.h +++ b/src/vnet/session/session.h @@ -15,13 +15,17 @@ #ifndef __included_session_h__ #define __included_session_h__ -#include -#include +#include +#include +#include #include #include +#include #define HALF_OPEN_LOOKUP_INVALID_VALUE ((u64)~0) #define INVALID_INDEX ((u32)~0) +#define SESSION_PROXY_LISTENER_INDEX ((u8)~0 - 1) +#define SESSION_LOCAL_HANDLE_PREFIX 0x7FFFFFFF /* TODO decide how much since we have pre-data as well */ #define MAX_HDRS_LEN 100 /* Max number of bytes for headers */ @@ -29,12 +33,50 @@ typedef enum { FIFO_EVENT_APP_RX, + SESSION_IO_EVT_CT_RX, FIFO_EVENT_APP_TX, - FIFO_EVENT_TIMEOUT, + SESSION_IO_EVT_CT_TX, FIFO_EVENT_DISCONNECT, FIFO_EVENT_BUILTIN_RX, + FIFO_EVENT_BUILTIN_TX, FIFO_EVENT_RPC, -} fifo_event_type_t; + SESSION_CTRL_EVT_BOUND, + SESSION_CTRL_EVT_ACCEPTED, + SESSION_CTRL_EVT_ACCEPTED_REPLY, + SESSION_CTRL_EVT_CONNECTED, + SESSION_CTRL_EVT_CONNECTED_REPLY, + SESSION_CTRL_EVT_DISCONNECTED, + SESSION_CTRL_EVT_DISCONNECTED_REPLY, + SESSION_CTRL_EVT_RESET, + SESSION_CTRL_EVT_RESET_REPLY +} session_evt_type_t; + +static inline const char * +fifo_event_type_str (session_evt_type_t et) +{ + switch (et) + { + case FIFO_EVENT_APP_RX: + return "FIFO_EVENT_APP_RX"; + case FIFO_EVENT_APP_TX: + return "FIFO_EVENT_APP_TX"; + case FIFO_EVENT_DISCONNECT: + return "FIFO_EVENT_DISCONNECT"; + case FIFO_EVENT_BUILTIN_RX: + return "FIFO_EVENT_BUILTIN_RX"; + case FIFO_EVENT_RPC: + return "FIFO_EVENT_RPC"; + default: + return "UNKNOWN FIFO EVENT"; + } +} + +typedef enum +{ + SESSION_MQ_IO_EVT_RING, + SESSION_MQ_CTRL_EVT_RING, + SESSION_MQ_N_RINGS +} session_mq_rings_e; #define foreach_session_input_error \ _(NO_SESSION, "No session drops") \ @@ -45,7 +87,8 @@ _(FIFO_FULL, "Packets dropped for lack of rx fifo space") \ _(EVENT_FIFO_FULL, "Events not sent for lack of event fifo space") \ _(API_QUEUE_FULL, "Sessions not created for lack of API queue space") \ _(NEW_SEG_NO_SPACE, "Created segment, couldn't allocate a fifo pair") \ -_(NO_SPACE, "Couldn't allocate a fifo pair") +_(NO_SPACE, "Couldn't allocate a fifo pair") \ +_(SEG_CREATE, "Couldn't create a new segment") typedef enum { @@ -55,169 +98,194 @@ typedef enum SESSION_N_ERROR, } session_error_t; -/* Event queue input node static next indices */ -typedef enum -{ - SESSION_QUEUE_NEXT_DROP, - SESSION_QUEUE_NEXT_TCP_IP4_OUTPUT, - SESSION_QUEUE_NEXT_IP4_LOOKUP, - SESSION_QUEUE_NEXT_TCP_IP6_OUTPUT, - SESSION_QUEUE_NEXT_IP6_LOOKUP, - SESSION_QUEUE_N_NEXT, -} session_queue_next_t; - -#define foreach_session_type \ - _(IP4_TCP, ip4_tcp) \ - _(IP4_UDP, ip4_udp) \ - _(IP6_TCP, ip6_tcp) \ - _(IP6_UDP, ip6_udp) - -typedef enum -{ -#define _(A, a) SESSION_TYPE_##A, - foreach_session_type -#undef _ - SESSION_N_TYPES, -} session_type_t; - -/* - * Application session state - */ -typedef enum -{ - SESSION_STATE_LISTENING, - SESSION_STATE_CONNECTING, - SESSION_STATE_READY, - SESSION_STATE_CLOSED, - SESSION_STATE_N_STATES, -} stream_session_state_t; - typedef struct { void *fp; void *arg; -} rpc_args_t; +} session_rpc_args_t; + +typedef u64 session_handle_t; /* *INDENT-OFF* */ -typedef CLIB_PACKED (struct { +typedef struct +{ + u8 event_type; + u8 postponed; union + { + svm_fifo_t *fifo; + session_handle_t session_handle; + session_rpc_args_t rpc_args; + struct { - svm_fifo_t * fifo; - u64 session_handle; - rpc_args_t rpc_args; + u8 data[0]; }; - u8 event_type; - u16 event_id; -}) session_fifo_event_t; + }; +} __clib_packed session_event_t; /* *INDENT-ON* */ -typedef struct _stream_session_t -{ - /** fifo pointers. Once allocated, these do not move */ - svm_fifo_t *server_rx_fifo; - svm_fifo_t *server_tx_fifo; - - /** Type */ - u8 session_type; - - /** State */ - u8 session_state; - - u8 thread_index; - - /** To avoid n**2 "one event per frame" check */ - u8 enqueue_epoch; - - /** Pad to a multiple of 8 octets */ - u8 align_pad[4]; - - /** svm segment index where fifos were allocated */ - u32 svm_segment_index; +#define SESSION_MSG_NULL { } - /** Session index in per_thread pool */ - u32 session_index; - - /** Transport specific */ - u32 connection_index; +typedef struct session_dgram_pre_hdr_ +{ + u32 data_length; + u32 data_offset; +} session_dgram_pre_hdr_t; - /** Application specific */ - u32 pid; +/* *INDENT-OFF* */ +typedef CLIB_PACKED (struct session_dgram_header_ +{ + u32 data_length; + u32 data_offset; + ip46_address_t rmt_ip; + ip46_address_t lcl_ip; + u16 rmt_port; + u16 lcl_port; + u8 is_ip4; +}) session_dgram_hdr_t; +/* *INDENT-ON* */ - /** stream server pool index */ - u32 app_index; +#define SESSION_CONN_ID_LEN 37 +#define SESSION_CONN_HDR_LEN 45 - /** Parent listener session if the result of an accept */ - u32 listener_index; +STATIC_ASSERT (sizeof (session_dgram_hdr_t) == (SESSION_CONN_ID_LEN + 8), + "session conn id wrong length"); - /** Opaque, pad to a 64-octet boundary */ - u64 opaque[2]; -} stream_session_t; +typedef struct session_tx_context_ +{ + CLIB_CACHE_LINE_ALIGN_MARK (cacheline0); + stream_session_t *s; + transport_proto_vft_t *transport_vft; + transport_connection_t *tc; + vlib_buffer_t *b; + u32 max_dequeue; + u32 snd_space; + u32 left_to_snd; + u32 tx_offset; + u32 max_len_to_snd; + u16 deq_per_first_buf; + u16 deq_per_buf; + u16 snd_mss; + u16 n_segs_per_evt; + u8 n_bufs_per_seg; + CLIB_CACHE_LINE_ALIGN_MARK (cacheline1); + session_dgram_hdr_t hdr; +} session_tx_context_t; /* Forward definition */ typedef struct _session_manager_main session_manager_main_t; typedef int (session_fifo_rx_fn) (vlib_main_t * vm, vlib_node_runtime_t * node, - session_manager_main_t * smm, - session_fifo_event_t * e0, stream_session_t * s0, - u32 thread_index, int *n_tx_pkts); + session_event_t * e0, stream_session_t * s0, + int *n_tx_pkts); extern session_fifo_rx_fn session_tx_fifo_peek_and_snd; extern session_fifo_rx_fn session_tx_fifo_dequeue_and_snd; +extern session_fifo_rx_fn session_tx_fifo_dequeue_internal; -struct _session_manager_main +u8 session_node_lookup_fifo_event (svm_fifo_t * f, session_event_t * e); + +typedef struct session_manager_worker_ { - /** Lookup tables for established sessions and listeners */ - clib_bihash_16_8_t v4_session_hash; - clib_bihash_48_8_t v6_session_hash; + CLIB_CACHE_LINE_ALIGN_MARK (cacheline0); - /** Lookup tables for half-open sessions */ - clib_bihash_16_8_t v4_half_open_hash; - clib_bihash_48_8_t v6_half_open_hash; + /** Worker session pool */ + stream_session_t *sessions; - /** Per worker thread session pools */ - stream_session_t **sessions; + /** vpp event message queue for worker */ + svm_msg_q_t *vpp_event_queue; - /** Pool of listen sessions. Same type as stream sessions to ease lookups */ - stream_session_t *listen_sessions[SESSION_N_TYPES]; + /** Our approximation of a "complete" dispatch loop period */ + f64 dispatch_period; - /** Sparse vector to map dst port to stream server */ - u16 *stream_server_by_dst_port[SESSION_N_TYPES]; + /** vlib_time_now last time around the track */ + f64 last_vlib_time; - /** per-worker enqueue epoch counters */ - u8 *current_enqueue_epoch; + /** Per-proto enqueue epoch counters */ + u64 current_enqueue_epoch[TRANSPORT_N_PROTO]; - /** Per-worker thread vector of sessions to enqueue */ - u32 **session_indices_to_enqueue_by_thread; + /** Per-proto vector of sessions to enqueue */ + u32 *session_to_enqueue[TRANSPORT_N_PROTO]; - /** per-worker tx buffer free lists */ - u32 **tx_buffers; + /** Context for session tx */ + session_tx_context_t ctx; - /** Per worker-thread vector of partially read events */ - session_fifo_event_t **free_event_vector; + /** Vector of tx buffer free lists */ + u32 *tx_buffers; - /** per-worker active event vectors */ - session_fifo_event_t **pending_event_vector; + /** Vector of partially read events */ + session_event_t *free_event_vector; - /** vpp fifo event queue */ - unix_shared_memory_queue_t **vpp_event_queues; + /** Vector of active event vectors */ + session_event_t *pending_event_vector; - /** vpp fifo event queue configured length */ - u32 configured_event_queue_length; + /** Vector of postponed disconnects */ + session_event_t *pending_disconnects; + + /** Vector of postponed events */ + session_event_t *postponed_event_vector; + + /** Peekers rw lock */ + clib_rwlock_t peekers_rw_locks; + + u32 last_tx_packets; + +} session_manager_worker_t; + +struct _session_manager_main +{ + /** Worker contexts */ + session_manager_worker_t *wrk; + + /** Event queues memfd segment initialized only if so configured */ + ssvm_private_t evt_qs_segment; /** Unique segment name counter */ u32 unique_segment_name_counter; /** Per transport rx function that can either dequeue or peek */ - session_fifo_rx_fn *session_tx_fns[SESSION_N_TYPES]; + session_fifo_rx_fn **session_tx_fns; + /** Per session type output nodes. Could optimize to group nodes by + * fib but lookup would then require session type parsing in session node. + * Trade memory for speed, for now */ + u32 *session_type_to_next; + + /* + * Config parameters + */ + + /** Session manager is enabled */ u8 is_enabled; - /* Convenience */ - vlib_main_t *vlib_main; - vnet_main_t *vnet_main; + /** vpp fifo event queue configured length */ + u32 configured_event_queue_length; -#if SESSION_DBG + /** Session ssvm segment configs*/ + uword session_baseva; + uword session_va_space_size; + u32 evt_qs_segment_size; + u8 evt_qs_use_memfd_seg; + + /** Session table size parameters */ + u32 configured_v4_session_table_buckets; + u32 configured_v4_session_table_memory; + u32 configured_v4_halfopen_table_buckets; + u32 configured_v4_halfopen_table_memory; + u32 configured_v6_session_table_buckets; + u32 configured_v6_session_table_memory; + u32 configured_v6_halfopen_table_buckets; + u32 configured_v6_halfopen_table_memory; + + /** Transport table (preallocation) size parameters */ + u32 local_endpoints_table_memory; + u32 local_endpoints_table_buckets; + + /** Preallocate session config parameter */ + u32 preallocated_sessions; + +#if SESSION_DEBUG /** * last event poll time by thread * Debug only. Will cause false cache-line sharing as-is @@ -229,6 +297,10 @@ struct _session_manager_main extern session_manager_main_t session_manager_main; extern vlib_node_registration_t session_queue_node; +extern vlib_node_registration_t session_queue_process_node; + +#define SESSION_Q_PROCESS_FLUSH_FRAMES 1 +#define SESSION_Q_PROCESS_STOP 2 /* * Session manager function @@ -239,137 +311,294 @@ vnet_get_session_manager_main () return &session_manager_main; } -/* - * Stream session functions - */ - -stream_session_t *stream_session_lookup_listener4 (ip4_address_t * lcl, - u16 lcl_port, u8 proto); -stream_session_t *stream_session_lookup4 (ip4_address_t * lcl, - ip4_address_t * rmt, u16 lcl_port, - u16 rmt_port, u8 proto, - u32 thread_index); -stream_session_t *stream_session_lookup_listener6 (ip6_address_t * lcl, - u16 lcl_port, u8 proto); -stream_session_t *stream_session_lookup6 (ip6_address_t * lcl, - ip6_address_t * rmt, u16 lcl_port, - u16 rmt_port, u8, u32 thread_index); -transport_connection_t - * stream_session_lookup_transport4 (ip4_address_t * lcl, - ip4_address_t * rmt, u16 lcl_port, - u16 rmt_port, u8 proto, - u32 thread_index); -transport_connection_t - * stream_session_lookup_transport6 (ip6_address_t * lcl, - ip6_address_t * rmt, u16 lcl_port, - u16 rmt_port, u8 proto, - u32 thread_index); -stream_session_t *stream_session_lookup_listener (ip46_address_t * lcl, - u16 lcl_port, u8 proto); -void stream_session_table_add_for_tc (transport_connection_t * tc, u64 value); -int stream_session_table_del_for_tc (transport_connection_t * tc); +always_inline session_manager_worker_t * +session_manager_get_worker (u32 thread_index) +{ + return &session_manager_main.wrk[thread_index]; +} -always_inline stream_session_t * -stream_session_get_tsi (u64 ti_and_si, u32 thread_index) +always_inline u8 +stream_session_is_valid (u32 si, u8 thread_index) { - ASSERT ((u32) (ti_and_si >> 32) == thread_index); - return pool_elt_at_index (session_manager_main.sessions[thread_index], - ti_and_si & 0xFFFFFFFFULL); + stream_session_t *s; + s = pool_elt_at_index (session_manager_main.wrk[thread_index].sessions, si); + if (s->thread_index != thread_index || s->session_index != si + /* || s->server_rx_fifo->master_session_index != si + || s->server_tx_fifo->master_session_index != si + || s->server_rx_fifo->master_thread_index != thread_index + || s->server_tx_fifo->master_thread_index != thread_index */ ) + return 0; + return 1; } +stream_session_t *session_alloc (u32 thread_index); +int session_alloc_fifos (segment_manager_t * sm, stream_session_t * s); +void session_free (stream_session_t * s); +void session_free_w_fifos (stream_session_t * s); + always_inline stream_session_t * -stream_session_get (u32 si, u32 thread_index) +session_get (u32 si, u32 thread_index) { - return pool_elt_at_index (session_manager_main.sessions[thread_index], si); + ASSERT (stream_session_is_valid (si, thread_index)); + return pool_elt_at_index (session_manager_main.wrk[thread_index].sessions, + si); } always_inline stream_session_t * -stream_session_get_if_valid (u64 si, u32 thread_index) +session_get_if_valid (u64 si, u32 thread_index) { - if (thread_index >= vec_len (session_manager_main.sessions)) + if (thread_index >= vec_len (session_manager_main.wrk)) return 0; - if (pool_is_free_index (session_manager_main.sessions[thread_index], si)) + if (pool_is_free_index (session_manager_main.wrk[thread_index].sessions, + si)) return 0; - return pool_elt_at_index (session_manager_main.sessions[thread_index], si); + ASSERT (stream_session_is_valid (si, thread_index)); + return pool_elt_at_index (session_manager_main.wrk[thread_index].sessions, + si); } -always_inline u64 -stream_session_handle (stream_session_t * s) +always_inline session_handle_t +session_handle (stream_session_t * s) { return ((u64) s->thread_index << 32) | (u64) s->session_index; } always_inline u32 -stream_session_index_from_handle (u64 handle) +session_index_from_handle (session_handle_t handle) { return handle & 0xFFFFFFFF; } always_inline u32 -stream_session_thread_from_handle (u64 handle) +session_thread_from_handle (session_handle_t handle) { return handle >> 32; } always_inline void -stream_session_parse_handle (u64 handle, u32 * index, u32 * thread_index) +session_parse_handle (session_handle_t handle, u32 * index, + u32 * thread_index) { - *index = stream_session_index_from_handle (handle); - *thread_index = stream_session_thread_from_handle (handle); + *index = session_index_from_handle (handle); + *thread_index = session_thread_from_handle (handle); } always_inline stream_session_t * -stream_session_get_from_handle (u64 handle) +session_get_from_handle (session_handle_t handle) { session_manager_main_t *smm = &session_manager_main; - return pool_elt_at_index (smm->sessions[stream_session_thread_from_handle - (handle)], - stream_session_index_from_handle (handle)); + u32 session_index, thread_index; + session_parse_handle (handle, &session_index, &thread_index); + return pool_elt_at_index (smm->wrk[thread_index].sessions, session_index); } always_inline stream_session_t * -stream_session_listener_get (u8 sst, u64 si) +session_get_from_handle_if_valid (session_handle_t handle) { - return pool_elt_at_index (session_manager_main.listen_sessions[sst], si); + u32 session_index, thread_index; + session_parse_handle (handle, &session_index, &thread_index); + return session_get_if_valid (session_index, thread_index); } -always_inline u32 -stream_session_get_index (stream_session_t * s) +always_inline u8 +session_handle_is_local (session_handle_t handle) +{ + if ((handle >> 32) == SESSION_LOCAL_HANDLE_PREFIX) + return 1; + return 0; +} + +always_inline transport_proto_t +session_type_transport_proto (session_type_t st) +{ + return (st >> 1); +} + +always_inline u8 +session_type_is_ip4 (session_type_t st) +{ + return (st & 1); +} + +always_inline transport_proto_t +session_get_transport_proto (stream_session_t * s) +{ + return (s->session_type >> 1); +} + +always_inline fib_protocol_t +session_get_fib_proto (stream_session_t * s) { - if (s->session_state == SESSION_STATE_LISTENING) - return s - session_manager_main.listen_sessions[s->session_type]; + u8 is_ip4 = s->session_type & 1; + return (is_ip4 ? FIB_PROTOCOL_IP4 : FIB_PROTOCOL_IP6); +} + +always_inline session_type_t +session_type_from_proto_and_ip (transport_proto_t proto, u8 is_ip4) +{ + return (proto << 1 | is_ip4); +} - return s - session_manager_main.sessions[s->thread_index]; +always_inline u64 +session_segment_handle (stream_session_t * s) +{ + svm_fifo_t *f = s->server_rx_fifo; + return segment_manager_make_segment_handle (f->segment_manager, + f->segment_index); +} + +always_inline u8 +session_has_transport (stream_session_t * s) +{ + return (session_get_transport_proto (s) != TRANSPORT_PROTO_NONE); +} + +transport_service_type_t session_transport_service_type (stream_session_t *); +transport_tx_fn_type_t session_transport_tx_fn_type (stream_session_t *); +u8 session_tx_is_dgram (stream_session_t * s); + +/** + * Acquires a lock that blocks a session pool from expanding. + * + * This is typically used for safely peeking into other threads' + * pools in order to clone elements. Lock should be dropped as soon + * as possible by calling @ref session_pool_remove_peeker. + * + * NOTE: Avoid using pool_elt_at_index while the lock is held because + * it may lead to free elt bitmap expansion/contraction! + */ +always_inline void +session_pool_add_peeker (u32 thread_index) +{ + session_manager_worker_t *wrk = &session_manager_main.wrk[thread_index]; + if (thread_index == vlib_get_thread_index ()) + return; + clib_rwlock_reader_lock (&wrk->peekers_rw_locks); +} + +always_inline void +session_pool_remove_peeker (u32 thread_index) +{ + session_manager_worker_t *wrk = &session_manager_main.wrk[thread_index]; + if (thread_index == vlib_get_thread_index ()) + return; + clib_rwlock_reader_unlock (&wrk->peekers_rw_locks); +} + +/** + * Get session from handle and 'lock' pool resize if not in same thread + * + * Caller should drop the peek 'lock' as soon as possible. + */ +always_inline stream_session_t * +session_get_from_handle_safe (u64 handle) +{ + u32 thread_index = session_thread_from_handle (handle); + session_manager_worker_t *wrk = &session_manager_main.wrk[thread_index]; + + if (thread_index == vlib_get_thread_index ()) + { + return pool_elt_at_index (wrk->sessions, + session_index_from_handle (handle)); + } + else + { + session_pool_add_peeker (thread_index); + /* Don't use pool_elt_at index. See @ref session_pool_add_peeker */ + return wrk->sessions + session_index_from_handle (handle); + } } always_inline u32 -stream_session_max_rx_enqueue (transport_connection_t * tc) +transport_max_rx_enqueue (transport_connection_t * tc) { - stream_session_t *s = stream_session_get (tc->s_index, tc->thread_index); + stream_session_t *s = session_get (tc->s_index, tc->thread_index); return svm_fifo_max_enqueue (s->server_rx_fifo); } always_inline u32 -stream_session_rx_fifo_size (transport_connection_t * tc) +transport_rx_fifo_size (transport_connection_t * tc) { - stream_session_t *s = stream_session_get (tc->s_index, tc->thread_index); + stream_session_t *s = session_get (tc->s_index, tc->thread_index); return s->server_rx_fifo->nitems; } -u32 stream_session_tx_fifo_max_dequeue (transport_connection_t * tc); +always_inline u32 +transport_tx_fifo_size (transport_connection_t * tc) +{ + stream_session_t *s = session_get (tc->s_index, tc->thread_index); + return s->server_tx_fifo->nitems; +} + +always_inline u8 +transport_rx_fifo_has_ooo_data (transport_connection_t * tc) +{ + stream_session_t *s = session_get (tc->c_index, tc->thread_index); + return svm_fifo_has_ooo_data (s->server_rx_fifo); +} + +always_inline f64 +transport_dispatch_period (u32 thread_index) +{ + return session_manager_main.wrk[thread_index].dispatch_period; +} + +always_inline f64 +transport_time_now (u32 thread_index) +{ + return session_manager_main.wrk[thread_index].last_vlib_time; +} + +always_inline u32 +session_get_index (stream_session_t * s) +{ + return (s - session_manager_main.wrk[s->thread_index].sessions); +} + +always_inline stream_session_t * +session_clone_safe (u32 session_index, u32 thread_index) +{ + stream_session_t *old_s, *new_s; + u32 current_thread_index = vlib_get_thread_index (); + + /* If during the memcpy pool is reallocated AND the memory allocator + * decides to give the old chunk of memory to somebody in a hurry to + * scribble something on it, we have a problem. So add this thread as + * a session pool peeker. + */ + session_pool_add_peeker (thread_index); + new_s = session_alloc (current_thread_index); + old_s = session_manager_main.wrk[thread_index].sessions + session_index; + clib_memcpy_fast (new_s, old_s, sizeof (*new_s)); + session_pool_remove_peeker (thread_index); + new_s->thread_index = current_thread_index; + new_s->session_index = session_get_index (new_s); + return new_s; +} + +transport_connection_t *session_get_transport (stream_session_t * s); + +u32 session_tx_fifo_max_dequeue (transport_connection_t * tc); int -stream_session_enqueue_data (transport_connection_t * tc, vlib_buffer_t * b, - u32 offset, u8 queue_event, u8 is_in_order); -int -stream_session_peek_bytes (transport_connection_t * tc, u8 * buffer, - u32 offset, u32 max_bytes); +session_enqueue_stream_connection (transport_connection_t * tc, + vlib_buffer_t * b, u32 offset, + u8 queue_event, u8 is_in_order); +int session_enqueue_dgram_connection (stream_session_t * s, + session_dgram_hdr_t * hdr, + vlib_buffer_t * b, u8 proto, + u8 queue_event); +int stream_session_peek_bytes (transport_connection_t * tc, u8 * buffer, + u32 offset, u32 max_bytes); u32 stream_session_dequeue_drop (transport_connection_t * tc, u32 max_bytes); -int stream_session_connect_notify (transport_connection_t * tc, u8 sst, - u8 is_fail); +int session_stream_connect_notify (transport_connection_t * tc, u8 is_fail); +int session_dgram_connect_notify (transport_connection_t * tc, + u32 old_thread_index, + stream_session_t ** new_session); +int session_dequeue_notify (stream_session_t * s); void stream_session_init_fifos_pointers (transport_connection_t * tc, u32 rx_pointer, u32 tx_pointer); @@ -377,97 +606,116 @@ void stream_session_accept_notify (transport_connection_t * tc); void stream_session_disconnect_notify (transport_connection_t * tc); void stream_session_delete_notify (transport_connection_t * tc); void stream_session_reset_notify (transport_connection_t * tc); -int -stream_session_accept (transport_connection_t * tc, u32 listener_index, - u8 sst, u8 notify); -int -stream_session_open (u32 app_index, session_type_t st, - transport_endpoint_t * tep, - transport_connection_t ** tc); -int stream_session_listen (stream_session_t * s, transport_endpoint_t * tep); -int stream_session_stop_listen (stream_session_t * s); +int stream_session_accept (transport_connection_t * tc, u32 listener_index, + u8 notify); +int session_open (u32 app_index, session_endpoint_t * tep, u32 opaque); +int session_listen (stream_session_t * s, session_endpoint_cfg_t * sep); +int session_stop_listen (stream_session_t * s); void stream_session_disconnect (stream_session_t * s); +void stream_session_disconnect_transport (stream_session_t * s); void stream_session_cleanup (stream_session_t * s); -void session_send_session_evt_to_thread (u64 session_handle, - fifo_event_type_t evt_type, - u32 thread_index); +int session_send_io_evt_to_thread (svm_fifo_t * f, + session_evt_type_t evt_type); +int session_send_io_evt_to_thread_custom (void *data, u32 thread_index, + session_evt_type_t evt_type); +void session_send_rpc_evt_to_thread (u32 thread_index, void *fp, + void *rpc_args); + +ssvm_private_t *session_manager_get_evt_q_segment (void); + u8 *format_stream_session (u8 * s, va_list * args); -int -send_session_connected_callback (u32 app_index, u32 api_context, - stream_session_t * s, u8 is_fail); +uword unformat_stream_session (unformat_input_t * input, va_list * args); +uword unformat_transport_connection (unformat_input_t * input, + va_list * args); +void session_register_transport (transport_proto_t transport_proto, + const transport_proto_vft_t * vft, u8 is_ip4, + u32 output_node); -void session_register_transport (u8 type, const transport_proto_vft_t * vft); -transport_proto_vft_t *session_get_transport_vft (u8 type); +always_inline void +transport_add_tx_event (transport_connection_t * tc) +{ + stream_session_t *s = session_get (tc->s_index, tc->thread_index); + if (svm_fifo_has_event (s->server_tx_fifo)) + return; + session_send_io_evt_to_thread (s->server_tx_fifo, FIFO_EVENT_APP_TX); +} clib_error_t *vnet_session_enable_disable (vlib_main_t * vm, u8 is_en); -always_inline unix_shared_memory_queue_t * +always_inline svm_msg_q_t * session_manager_get_vpp_event_queue (u32 thread_index) { - return session_manager_main.vpp_event_queues[thread_index]; + return session_manager_main.wrk[thread_index].vpp_event_queue; } -int session_manager_flush_enqueue_events (u32 thread_index); +int session_manager_flush_enqueue_events (u8 proto, u32 thread_index); +int session_manager_flush_all_enqueue_events (u8 transport_proto); always_inline u64 listen_session_get_handle (stream_session_t * s) { ASSERT (s->session_state == SESSION_STATE_LISTENING); - return ((u64) s->session_type << 32) | s->session_index; + return session_handle (s); } always_inline stream_session_t * -listen_session_get_from_handle (u64 handle) +listen_session_get_from_handle (session_handle_t handle) { - session_manager_main_t *smm = &session_manager_main; - stream_session_t *s; - u32 type, index; - type = handle >> 32; - index = handle & 0xFFFFFFFF; - - if (pool_is_free_index (smm->listen_sessions[type], index)) - return 0; + return session_get_from_handle (handle); +} - s = pool_elt_at_index (smm->listen_sessions[type], index); - ASSERT (s->session_state == SESSION_STATE_LISTENING); - return s; +always_inline void +listen_session_parse_handle (session_handle_t handle, u32 * index, + u32 * thread_index) +{ + session_parse_handle (handle, index, thread_index); } always_inline stream_session_t * -listen_session_new (session_type_t type) +listen_session_new (u8 thread_index, session_type_t type) { stream_session_t *s; - pool_get_aligned (session_manager_main.listen_sessions[type], s, - CLIB_CACHE_LINE_BYTES); - memset (s, 0, sizeof (*s)); - + s = session_alloc (thread_index); s->session_type = type; s->session_state = SESSION_STATE_LISTENING; - s->session_index = s - session_manager_main.listen_sessions[type]; - return s; } always_inline stream_session_t * -listen_session_get (session_type_t type, u32 index) +listen_session_get (u32 index) { - return pool_elt_at_index (session_manager_main.listen_sessions[type], - index); + return session_get (index, 0); } always_inline void listen_session_del (stream_session_t * s) { - pool_put (session_manager_main.listen_sessions[s->session_type], s); + session_free (s); } +transport_connection_t *listen_session_get_transport (stream_session_t * s); + +int +listen_session_get_local_session_endpoint (stream_session_t * listener, + session_endpoint_t * sep); + +void session_flush_frames_main_thread (vlib_main_t * vm); + always_inline u8 session_manager_is_enabled () { return session_manager_main.is_enabled == 1; } +#define session_cli_return_if_not_enabled() \ +do { \ + if (!session_manager_main.is_enabled) \ + return clib_error_return(0, "session layer is not enabled"); \ +} while (0) + +void session_node_enable_disable (u8 is_en); + #endif /* __included_session_h__ */ /*