#ifndef __included_session_h__
#define __included_session_h__
+#include <vppinfra/llist.h>
#include <vnet/session/session_types.h>
#include <vnet/session/session_lookup.h>
#include <vnet/session/session_debug.h>
-#include <vnet/session/segment_manager.h>
#include <svm/message_queue.h>
-
-#define SESSION_PROXY_LISTENER_INDEX ((u8)~0 - 1)
+#include <svm/fifo_segment.h>
#define foreach_session_input_error \
_(NO_SESSION, "No session drops") \
_(FIFO_FULL, "Packets dropped for lack of rx fifo space") \
_(EVENT_FIFO_FULL, "Events not sent for lack of event fifo space") \
_(API_QUEUE_FULL, "Sessions not created for lack of API queue space") \
-_(NEW_SEG_NO_SPACE, "Created segment, couldn't allocate a fifo pair") \
-_(NO_SPACE, "Couldn't allocate a fifo pair") \
-_(SEG_CREATE, "Couldn't create a new segment")
typedef enum
{
foreach_session_input_error
#undef _
SESSION_N_ERROR,
-} session_error_t;
+} session_input_error_t;
typedef struct session_tx_context_
{
session_t *s;
transport_proto_vft_t *transport_vft;
transport_connection_t *tc;
- vlib_buffer_t *b;
+ transport_send_params_t sp;
u32 max_dequeue;
- u32 snd_space;
u32 left_to_snd;
- u32 tx_offset;
u32 max_len_to_snd;
u16 deq_per_first_buf;
u16 deq_per_buf;
- u16 snd_mss;
u16 n_segs_per_evt;
+ u16 n_bufs_needed;
u8 n_bufs_per_seg;
CLIB_CACHE_LINE_ALIGN_MARK (cacheline1);
session_dgram_hdr_t hdr;
+
+ /** Vector of tx buffer free lists */
+ u32 *tx_buffers;
+ vlib_buffer_t **transport_pending_bufs;
} session_tx_context_t;
-typedef struct session_manager_worker_
+typedef struct session_evt_elt
+{
+ clib_llist_anchor_t evt_list;
+ session_event_t evt;
+} session_evt_elt_t;
+
+typedef struct session_ctrl_evt_data_
+{
+ u8 data[SESSION_CTRL_MSG_MAX_SIZE];
+} session_evt_ctrl_data_t;
+
+typedef enum session_wrk_state_
+{
+ SESSION_WRK_POLLING,
+ SESSION_WRK_INTERRUPT,
+ SESSION_WRK_IDLE,
+} __clib_packed session_wrk_state_t;
+
+typedef enum session_wrk_flags_
+{
+ SESSION_WRK_F_ADAPTIVE = 1 << 0,
+} __clib_packed session_wrk_flag_t;
+
+typedef struct session_worker_
{
CLIB_CACHE_LINE_ALIGN_MARK (cacheline0);
/** vpp event message queue for worker */
svm_msg_q_t *vpp_event_queue;
- /** Our approximation of a "complete" dispatch loop period */
- f64 dispatch_period;
-
/** vlib_time_now last time around the track */
- f64 last_vlib_time;
+ clib_time_type_t last_vlib_time;
+
+ /** vlib_time_now rounded to us precision and as u64 */
+ clib_us_time_t last_vlib_us_time;
- /** Per-proto enqueue epoch counters */
- u64 current_enqueue_epoch[TRANSPORT_N_PROTO];
+ /** Convenience pointer to this worker's vlib_main */
+ vlib_main_t *vm;
/** Per-proto vector of sessions to enqueue */
- u32 *session_to_enqueue[TRANSPORT_N_PROTO];
+ u32 **session_to_enqueue;
+
+ /** Timerfd used to periodically signal wrk session queue node */
+ int timerfd;
+
+ /** Worker flags */
+ session_wrk_flag_t flags;
+
+ /** Worker state */
+ session_wrk_state_t state;
/** Context for session tx */
session_tx_context_t ctx;
- /** Vector of tx buffer free lists */
- u32 *tx_buffers;
+ /** Pool of session event list elements */
+ session_evt_elt_t *event_elts;
+
+ /** Pool of ctrl events data buffers */
+ session_evt_ctrl_data_t *ctrl_evts_data;
+
+ /** Head of control events list */
+ clib_llist_index_t ctrl_head;
+
+ /** Head of list of elements */
+ clib_llist_index_t new_head;
+
+ /** Head of list of pending events */
+ clib_llist_index_t old_head;
- /** Vector of partially read events */
- session_event_t *free_event_vector;
+ /** Vector of buffers to be sent */
+ u32 *pending_tx_buffers;
- /** Vector of active event vectors */
- session_event_t *pending_event_vector;
+ /** Vector of nexts for the pending tx buffers */
+ u16 *pending_tx_nexts;
- /** Vector of postponed disconnects */
- session_event_t *pending_disconnects;
+ /** Clib file for timerfd. Used only if adaptive mode is on */
+ uword timerfd_file;
- /** Vector of postponed events */
- session_event_t *postponed_event_vector;
+ /** List of pending connects for first worker */
+ clib_llist_index_t pending_connects;
- /** Peekers rw lock */
- clib_rwlock_t peekers_rw_locks;
+ /** Flag that is set if main thread signaled to handle connects */
+ u32 n_pending_connects;
- u32 last_tx_packets;
+ /** Main thread loops in poll mode without a connect */
+ u32 no_connect_loops;
-} session_manager_worker_t;
+ /** List head for first worker evts pending handling on main */
+ clib_llist_index_t evts_pending_main;
-typedef int (session_fifo_rx_fn) (vlib_main_t * vm,
+#if SESSION_DEBUG
+ /** last event poll time by thread */
+ clib_time_type_t last_event_poll;
+#endif
+} session_worker_t;
+
+typedef int (session_fifo_rx_fn) (session_worker_t * wrk,
vlib_node_runtime_t * node,
- session_manager_worker_t * wrk,
- session_event_t * e, int *n_tx_pkts);
+ session_evt_elt_t * e, int *n_tx_packets);
extern session_fifo_rx_fn session_tx_fifo_peek_and_snd;
extern session_fifo_rx_fn session_tx_fifo_dequeue_and_snd;
u8 session_node_lookup_fifo_event (svm_fifo_t * f, session_event_t * e);
-typedef struct session_manager_main_
+typedef void (*session_update_time_fn) (f64 time_now, u8 thread_index);
+
+typedef struct session_main_
{
/** Worker contexts */
- session_manager_worker_t *wrk;
+ session_worker_t *wrk;
+
+ /** Vector of transport update time functions */
+ session_update_time_fn *update_time_fns;
- /** Event queues memfd segment initialized only if so configured */
- ssvm_private_t evt_qs_segment;
+ /** Event queues memfd segment */
+ fifo_segment_t wrk_mqs_segment;
/** Unique segment name counter */
u32 unique_segment_name_counter;
* Trade memory for speed, for now */
u32 *session_type_to_next;
+ /** Thread for cl and ho that rely on cl allocs */
+ u32 transport_cl_thread;
+
+ transport_proto_t last_transport_proto_type;
+
+ /** Number of workers at pool realloc barrier */
+ u32 pool_realloc_at_barrier;
+
+ /** Lock to synchronize parallel forced reallocs */
+ clib_spinlock_t pool_realloc_lock;
+
/*
* Config parameters
*/
/** Session manager is enabled */
u8 is_enabled;
+ /** Session manager initialized (not necessarily enabled) */
+ u8 is_initialized;
+
+ /** Enable session manager at startup */
+ u8 session_enable_asap;
+
+ /** Poll session node in main thread */
+ u8 poll_main;
+
+ /** Allocate private rx mqs for external apps */
+ u8 use_private_rx_mqs;
+
+ /** Do not enable session queue node adaptive mode */
+ u8 no_adaptive;
+
/** vpp fifo event queue configured length */
- u32 configured_event_queue_length;
+ u32 configured_wrk_mq_length;
/** Session ssvm segment configs*/
- uword session_baseva;
- uword session_va_space_size;
- uword evt_qs_segment_size;
- u8 evt_qs_use_memfd_seg;
+ uword wrk_mqs_segment_size;
/** Session table size parameters */
u32 configured_v4_session_table_buckets;
/** Preallocate session config parameter */
u32 preallocated_sessions;
-#if SESSION_DEBUG
- /**
- * last event poll time by thread
- * Debug only. Will cause false cache-line sharing as-is
- */
- f64 *last_event_poll_by_thread;
-#endif
+ u16 msg_id_base;
+} session_main_t;
-} session_manager_main_t;
-
-extern session_manager_main_t session_manager_main;
+extern session_main_t session_main;
extern vlib_node_registration_t session_queue_node;
extern vlib_node_registration_t session_queue_process_node;
+extern vlib_node_registration_t session_queue_pre_input_node;
-#define SESSION_Q_PROCESS_FLUSH_FRAMES 1
-#define SESSION_Q_PROCESS_STOP 2
+typedef enum session_q_process_evt_
+{
+ SESSION_Q_PROCESS_RUN_ON_MAIN = 1,
+ SESSION_Q_PROCESS_STOP
+} session_q_process_evt_t;
-always_inline u8
-stream_session_is_valid (u32 si, u8 thread_index)
+#define TRANSPORT_PROTO_INVALID (session_main.last_transport_proto_type + 1)
+#define TRANSPORT_N_PROTOS (session_main.last_transport_proto_type + 1)
+
+static inline void
+session_evt_add_old (session_worker_t * wrk, session_evt_elt_t * elt)
{
- session_t *s;
- s = pool_elt_at_index (session_manager_main.wrk[thread_index].sessions, si);
- if (s->thread_index != thread_index || s->session_index != si
- /* || s->server_rx_fifo->master_session_index != si
- || s->server_tx_fifo->master_session_index != si
- || s->server_rx_fifo->master_thread_index != thread_index
- || s->server_tx_fifo->master_thread_index != thread_index */ )
- return 0;
- return 1;
+ clib_llist_add_tail (wrk->event_elts, evt_list, elt,
+ clib_llist_elt (wrk->event_elts, wrk->old_head));
+}
+
+static inline void
+session_evt_add_head_old (session_worker_t * wrk, session_evt_elt_t * elt)
+{
+ clib_llist_add (wrk->event_elts, evt_list, elt,
+ clib_llist_elt (wrk->event_elts, wrk->old_head));
+}
+
+
+static inline u32
+session_evt_ctrl_data_alloc (session_worker_t * wrk)
+{
+ session_evt_ctrl_data_t *data;
+ pool_get (wrk->ctrl_evts_data, data);
+ return (data - wrk->ctrl_evts_data);
+}
+
+static inline session_evt_elt_t *
+session_evt_alloc_ctrl (session_worker_t * wrk)
+{
+ session_evt_elt_t *elt;
+ clib_llist_get (wrk->event_elts, elt);
+ clib_llist_add_tail (wrk->event_elts, evt_list, elt,
+ clib_llist_elt (wrk->event_elts, wrk->ctrl_head));
+ return elt;
+}
+
+static inline void *
+session_evt_ctrl_data (session_worker_t * wrk, session_evt_elt_t * elt)
+{
+ return (void *) (pool_elt_at_index (wrk->ctrl_evts_data,
+ elt->evt.ctrl_data_index));
+}
+
+static inline void
+session_evt_ctrl_data_free (session_worker_t * wrk, session_evt_elt_t * elt)
+{
+ ASSERT (elt->evt.event_type > SESSION_IO_EVT_BUILTIN_TX);
+ pool_put_index (wrk->ctrl_evts_data, elt->evt.ctrl_data_index);
+}
+
+static inline session_evt_elt_t *
+session_evt_alloc_new (session_worker_t * wrk)
+{
+ session_evt_elt_t *elt;
+ clib_llist_get (wrk->event_elts, elt);
+ clib_llist_add_tail (wrk->event_elts, evt_list, elt,
+ clib_llist_elt (wrk->event_elts, wrk->new_head));
+ return elt;
+}
+
+static inline session_evt_elt_t *
+session_evt_alloc_old (session_worker_t * wrk)
+{
+ session_evt_elt_t *elt;
+ clib_llist_get (wrk->event_elts, elt);
+ clib_llist_add_tail (wrk->event_elts, evt_list, elt,
+ clib_llist_elt (wrk->event_elts, wrk->old_head));
+ return elt;
}
+int session_wrk_handle_mq (session_worker_t *wrk, svm_msg_q_t *mq);
+
session_t *session_alloc (u32 thread_index);
void session_free (session_t * s);
void session_free_w_fifos (session_t * s);
+void session_cleanup_half_open (session_handle_t ho_handle);
+u8 session_is_valid (u32 si, u8 thread_index);
always_inline session_t *
session_get (u32 si, u32 thread_index)
{
- ASSERT (stream_session_is_valid (si, thread_index));
- return pool_elt_at_index (session_manager_main.wrk[thread_index].sessions,
- si);
+ ASSERT (session_is_valid (si, thread_index));
+ return pool_elt_at_index (session_main.wrk[thread_index].sessions, si);
}
always_inline session_t *
session_get_if_valid (u64 si, u32 thread_index)
{
- if (thread_index >= vec_len (session_manager_main.wrk))
+ if (thread_index >= vec_len (session_main.wrk))
return 0;
- if (pool_is_free_index (session_manager_main.wrk[thread_index].sessions,
- si))
+ if (pool_is_free_index (session_main.wrk[thread_index].sessions, si))
return 0;
- ASSERT (stream_session_is_valid (si, thread_index));
- return pool_elt_at_index (session_manager_main.wrk[thread_index].sessions,
- si);
+ ASSERT (session_is_valid (si, thread_index));
+ return pool_elt_at_index (session_main.wrk[thread_index].sessions, si);
}
always_inline session_t *
session_get_from_handle (session_handle_t handle)
{
- session_manager_main_t *smm = &session_manager_main;
+ session_main_t *smm = &session_main;
u32 session_index, thread_index;
session_parse_handle (handle, &session_index, &thread_index);
return pool_elt_at_index (smm->wrk[thread_index].sessions, session_index);
return session_get_if_valid (session_index, thread_index);
}
-always_inline u64
-session_segment_handle (session_t * s)
-{
- svm_fifo_t *f;
-
- if (s->session_state == SESSION_STATE_LISTENING)
- return SESSION_INVALID_HANDLE;
-
- f = s->rx_fifo;
- return segment_manager_make_segment_handle (f->segment_manager,
- f->segment_index);
-}
-
-/**
- * Acquires a lock that blocks a session pool from expanding.
- *
- * This is typically used for safely peeking into other threads'
- * pools in order to clone elements. Lock should be dropped as soon
- * as possible by calling @ref session_pool_remove_peeker.
- *
- * NOTE: Avoid using pool_elt_at_index while the lock is held because
- * it may lead to free elt bitmap expansion/contraction!
- */
-always_inline void
-session_pool_add_peeker (u32 thread_index)
-{
- session_manager_worker_t *wrk = &session_manager_main.wrk[thread_index];
- if (thread_index == vlib_get_thread_index ())
- return;
- clib_rwlock_reader_lock (&wrk->peekers_rw_locks);
-}
-
-always_inline void
-session_pool_remove_peeker (u32 thread_index)
-{
- session_manager_worker_t *wrk = &session_manager_main.wrk[thread_index];
- if (thread_index == vlib_get_thread_index ())
- return;
- clib_rwlock_reader_unlock (&wrk->peekers_rw_locks);
-}
+u64 session_segment_handle (session_t * s);
/**
- * Get session from handle and 'lock' pool resize if not in same thread
+ * Get session from handle and avoid pool validation if no same thread
*
- * Caller should drop the peek 'lock' as soon as possible.
+ * Peekers are fine because pool grows with barrier (see @ref session_alloc)
*/
always_inline session_t *
session_get_from_handle_safe (u64 handle)
{
u32 thread_index = session_thread_from_handle (handle);
- session_manager_worker_t *wrk = &session_manager_main.wrk[thread_index];
+ session_worker_t *wrk = &session_main.wrk[thread_index];
if (thread_index == vlib_get_thread_index ())
{
}
else
{
- session_pool_add_peeker (thread_index);
- /* Don't use pool_elt_at index. See @ref session_pool_add_peeker */
+ /* Don't use pool_elt_at index to avoid pool bitmap reallocs */
return wrk->sessions + session_index_from_handle (handle);
}
}
-always_inline u32
-session_get_index (session_t * s)
-{
- return (s - session_manager_main.wrk[s->thread_index].sessions);
-}
-
always_inline session_t *
session_clone_safe (u32 session_index, u32 thread_index)
{
+ u32 current_thread_index = vlib_get_thread_index (), new_index;
session_t *old_s, *new_s;
- u32 current_thread_index = vlib_get_thread_index ();
- /* If during the memcpy pool is reallocated AND the memory allocator
- * decides to give the old chunk of memory to somebody in a hurry to
- * scribble something on it, we have a problem. So add this thread as
- * a session pool peeker.
- */
- session_pool_add_peeker (thread_index);
new_s = session_alloc (current_thread_index);
- old_s = session_manager_main.wrk[thread_index].sessions + session_index;
+ new_index = new_s->session_index;
+ /* Session pools are reallocated with barrier (see @ref session_alloc) */
+ old_s = session_main.wrk[thread_index].sessions + session_index;
clib_memcpy_fast (new_s, old_s, sizeof (*new_s));
- session_pool_remove_peeker (thread_index);
new_s->thread_index = current_thread_index;
- new_s->session_index = session_get_index (new_s);
+ new_s->session_index = new_index;
return new_s;
}
-int session_open (u32 app_index, session_endpoint_t * tep, u32 opaque);
+int session_open (session_endpoint_cfg_t *sep, session_handle_t *rsh);
int session_listen (session_t * s, session_endpoint_cfg_t * sep);
int session_stop_listen (session_t * s);
+void session_half_close (session_t *s);
void session_close (session_t * s);
+void session_reset (session_t * s);
+void session_transport_half_close (session_t *s);
void session_transport_close (session_t * s);
+void session_transport_reset (session_t * s);
void session_transport_cleanup (session_t * s);
int session_send_io_evt_to_thread (svm_fifo_t * f,
session_evt_type_t evt_type);
+int session_enqueue_notify (session_t * s);
int session_dequeue_notify (session_t * s);
int session_send_io_evt_to_thread_custom (void *data, u32 thread_index,
session_evt_type_t evt_type);
void session_send_rpc_evt_to_thread (u32 thread_index, void *fp,
void *rpc_args);
+void session_send_rpc_evt_to_thread_force (u32 thread_index, void *fp,
+ void *rpc_args);
+void session_add_self_custom_tx_evt (transport_connection_t * tc,
+ u8 has_prio);
+void sesssion_reschedule_tx (transport_connection_t * tc);
transport_connection_t *session_get_transport (session_t * s);
+void session_get_endpoint (session_t * s, transport_endpoint_t * tep,
+ u8 is_lcl);
+int session_transport_attribute (session_t *s, u8 is_get,
+ transport_endpt_attr_t *attr);
-
-u8 *format_stream_session (u8 * s, va_list * args);
-uword unformat_stream_session (unformat_input_t * input, va_list * args);
+u8 *format_session (u8 * s, va_list * args);
+uword unformat_session (unformat_input_t * input, va_list * args);
uword unformat_transport_connection (unformat_input_t * input,
va_list * args);
session_dgram_hdr_t * hdr,
vlib_buffer_t * b, u8 proto,
u8 queue_event);
-int stream_session_peek_bytes (transport_connection_t * tc, u8 * buffer,
- u32 offset, u32 max_bytes);
-u32 session_dequeue_drop (transport_connection_t * tc, u32 max_bytes);
-
-int session_stream_connect_notify (transport_connection_t * tc, u8 is_fail);
+int session_stream_connect_notify (transport_connection_t * tc,
+ session_error_t err);
int session_dgram_connect_notify (transport_connection_t * tc,
u32 old_thread_index,
session_t ** new_session);
-void stream_session_init_fifos_pointers (transport_connection_t * tc,
- u32 rx_pointer, u32 tx_pointer);
-
int session_stream_accept_notify (transport_connection_t * tc);
void session_transport_closing_notify (transport_connection_t * tc);
void session_transport_delete_notify (transport_connection_t * tc);
+void session_half_open_delete_notify (transport_connection_t *tc);
+void session_half_open_migrate_notify (transport_connection_t *tc);
+int session_half_open_migrated_notify (transport_connection_t *tc);
void session_transport_closed_notify (transport_connection_t * tc);
void session_transport_reset_notify (transport_connection_t * tc);
int session_stream_accept (transport_connection_t * tc, u32 listener_index,
- u8 notify);
-u32 session_tx_fifo_max_dequeue (transport_connection_t * tc);
+ u32 thread_index, u8 notify);
+int session_dgram_accept (transport_connection_t * tc, u32 listener_index,
+ u32 thread_index);
+/**
+ * Initialize session layer for given transport proto and ip version
+ *
+ * Allocates per session type (transport proto + ip version) data structures
+ * and adds arc from session queue node to session type output node.
+ *
+ * @param transport_proto transport proto to be registered
+ * @param vft virtual function table for transport
+ * @param is_ip4 flag that indicates if transports uses ipv4
+ * as underlying network layer
+ * @param output_node output node for transport
+ */
void session_register_transport (transport_proto_t transport_proto,
const transport_proto_vft_t * vft, u8 is_ip4,
u32 output_node);
+transport_proto_t session_add_transport_proto (void);
+void session_register_update_time_fn (session_update_time_fn fn, u8 is_add);
+int session_tx_fifo_peek_bytes (transport_connection_t * tc, u8 * buffer,
+ u32 offset, u32 max_bytes);
+u32 session_tx_fifo_dequeue_drop (transport_connection_t * tc, u32 max_bytes);
always_inline u32
transport_max_rx_enqueue (transport_connection_t * tc)
{
session_t *s = session_get (tc->s_index, tc->thread_index);
- return svm_fifo_max_enqueue (s->rx_fifo);
+ return svm_fifo_max_enqueue_prod (s->rx_fifo);
}
always_inline u32
transport_max_tx_dequeue (transport_connection_t * tc)
{
session_t *s = session_get (tc->s_index, tc->thread_index);
- return svm_fifo_max_dequeue (s->tx_fifo);
+ return svm_fifo_max_dequeue_cons (s->tx_fifo);
+}
+
+always_inline u32
+transport_max_rx_dequeue (transport_connection_t * tc)
+{
+ session_t *s = session_get (tc->s_index, tc->thread_index);
+ return svm_fifo_max_dequeue (s->rx_fifo);
}
always_inline u32
transport_rx_fifo_size (transport_connection_t * tc)
{
session_t *s = session_get (tc->s_index, tc->thread_index);
- return s->rx_fifo->nitems;
+ return svm_fifo_size (s->rx_fifo);
}
always_inline u32
transport_tx_fifo_size (transport_connection_t * tc)
{
session_t *s = session_get (tc->s_index, tc->thread_index);
- return s->tx_fifo->nitems;
+ return svm_fifo_size (s->tx_fifo);
}
always_inline u8
return svm_fifo_has_ooo_data (s->rx_fifo);
}
-always_inline f64
-transport_dispatch_period (u32 thread_index)
+always_inline void
+transport_rx_fifo_req_deq_ntf (transport_connection_t *tc)
{
- return session_manager_main.wrk[thread_index].dispatch_period;
+ session_t *s = session_get (tc->s_index, tc->thread_index);
+ svm_fifo_add_want_deq_ntf (s->rx_fifo, SVM_FIFO_WANT_DEQ_NOTIF);
}
-always_inline f64
+always_inline clib_time_type_t
transport_time_now (u32 thread_index)
{
- return session_manager_main.wrk[thread_index].last_vlib_time;
+ return session_main.wrk[thread_index].last_vlib_time;
+}
+
+always_inline clib_us_time_t
+transport_us_time_now (u32 thread_index)
+{
+ return session_main.wrk[thread_index].last_vlib_us_time;
+}
+
+always_inline clib_time_type_t
+transport_seconds_per_loop (u32 thread_index)
+{
+ return session_main.wrk[thread_index].vm->seconds_per_loop;
}
always_inline void
session_t *s = session_get (tc->s_index, tc->thread_index);
if (svm_fifo_has_event (s->tx_fifo))
return;
- session_send_io_evt_to_thread (s->tx_fifo, FIFO_EVENT_APP_TX);
+ session_send_io_evt_to_thread (s->tx_fifo, SESSION_IO_EVT_TX);
+}
+
+always_inline u32
+transport_cl_thread (void)
+{
+ return session_main.transport_cl_thread;
}
/*
always_inline u64
listen_session_get_handle (session_t * s)
{
- ASSERT (s->session_state == SESSION_STATE_LISTENING);
+ ASSERT (s->session_state == SESSION_STATE_LISTENING ||
+ session_get_transport_proto (s) == TRANSPORT_PROTO_QUIC);
return session_handle (s);
}
}
always_inline session_t *
-listen_session_new (u8 thread_index, session_type_t type)
+listen_session_alloc (u8 thread_index, session_type_t type)
{
session_t *s;
s = session_alloc (thread_index);
}
always_inline session_t *
-listen_session_get (u32 index)
+listen_session_get (u32 ls_index)
{
- return session_get (index, 0);
+ return session_get (ls_index, 0);
}
always_inline void
-listen_session_del (session_t * s)
+listen_session_free (session_t * s)
{
+ ASSERT (!s->rx_fifo);
session_free (s);
}
-transport_connection_t *listen_session_get_transport (session_t * s);
+always_inline session_t *
+ho_session_alloc (void)
+{
+ session_t *s;
+ ASSERT (vlib_get_thread_index () == 0);
+ s = session_alloc (0);
+ s->session_state = SESSION_STATE_CONNECTING;
+ s->flags |= SESSION_F_HALF_OPEN;
+ /* Not ideal. Half-opens are only allocated from main with worker barrier
+ * but can be cleaned up, i.e., session_half_open_free, from main without
+ * a barrier. In debug images, the free_bitmap can grow while workers peek
+ * the sessions pool, e.g., session_half_open_migrate_notify, and as a
+ * result crash while validating the session. To avoid this, grow the bitmap
+ * now. */
+ if (CLIB_DEBUG)
+ {
+ session_t *sp = session_main.wrk[0].sessions;
+ clib_bitmap_validate (pool_header (sp)->free_bitmap, s->session_index);
+ }
+ return s;
+}
-int
-listen_session_get_local_session_endpoint (session_t * listener,
- session_endpoint_t * sep);
+always_inline session_t *
+ho_session_get (u32 ho_index)
+{
+ return session_get (ho_index, 0 /* half-open thread */);
+}
+
+always_inline void
+ho_session_free (session_t *s)
+{
+ ASSERT (!s->rx_fifo && s->thread_index == 0);
+ session_free (s);
+}
+
+transport_connection_t *listen_session_get_transport (session_t * s);
/*
- * Session manager functions
+ * Session layer functions
*/
-always_inline session_manager_main_t *
-vnet_get_session_manager_main ()
+always_inline session_main_t *
+vnet_get_session_main ()
{
- return &session_manager_main;
+ return &session_main;
}
-always_inline session_manager_worker_t *
-session_manager_get_worker (u32 thread_index)
+always_inline session_worker_t *
+session_main_get_worker (u32 thread_index)
{
- return &session_manager_main.wrk[thread_index];
+ return &session_main.wrk[thread_index];
+}
+
+static inline session_worker_t *
+session_main_get_worker_if_valid (u32 thread_index)
+{
+ if (thread_index > vec_len (session_main.wrk))
+ return 0;
+ return &session_main.wrk[thread_index];
}
always_inline svm_msg_q_t *
-session_manager_get_vpp_event_queue (u32 thread_index)
+session_main_get_vpp_event_queue (u32 thread_index)
{
- return session_manager_main.wrk[thread_index].vpp_event_queue;
+ return session_main.wrk[thread_index].vpp_event_queue;
}
always_inline u8
-session_manager_is_enabled ()
+session_main_is_enabled ()
{
- return session_manager_main.is_enabled == 1;
+ return session_main.is_enabled == 1;
}
#define session_cli_return_if_not_enabled() \
do { \
- if (!session_manager_main.is_enabled) \
- return clib_error_return(0, "session layer is not enabled"); \
+ if (!session_main.is_enabled) \
+ return clib_error_return (0, "session layer is not enabled"); \
} while (0)
-int session_manager_flush_enqueue_events (u8 proto, u32 thread_index);
-int session_manager_flush_all_enqueue_events (u8 transport_proto);
-void session_flush_frames_main_thread (vlib_main_t * vm);
-ssvm_private_t *session_manager_get_evt_q_segment (void);
+int session_main_flush_enqueue_events (u8 proto, u32 thread_index);
+int session_main_flush_all_enqueue_events (u8 transport_proto);
+void session_queue_run_on_main_thread (vlib_main_t * vm);
+
+/**
+ * Add session node pending buffer with custom node
+ *
+ * @param thread_index worker thread expected to send the buffer
+ * @param bi buffer index
+ * @param next_node next node edge index for buffer. Edge to next node
+ * must exist
+ */
+always_inline void
+session_add_pending_tx_buffer (u32 thread_index, u32 bi, u32 next_node)
+{
+ session_worker_t *wrk = session_main_get_worker (thread_index);
+ vec_add1 (wrk->pending_tx_buffers, bi);
+ vec_add1 (wrk->pending_tx_nexts, next_node);
+ if (PREDICT_FALSE (wrk->state == SESSION_WRK_INTERRUPT))
+ vlib_node_set_interrupt_pending (wrk->vm, session_queue_node.index);
+}
+
+always_inline void
+session_wrk_update_time (session_worker_t *wrk, f64 now)
+{
+ wrk->last_vlib_time = now;
+ wrk->last_vlib_us_time = wrk->last_vlib_time * CLIB_US_TIME_FREQ;
+}
+
+void session_wrk_enable_adaptive_mode (session_worker_t *wrk);
+fifo_segment_t *session_main_get_wrk_mqs_segment (void);
void session_node_enable_disable (u8 is_en);
clib_error_t *vnet_session_enable_disable (vlib_main_t * vm, u8 is_en);
+void session_wrk_handle_evts_main_rpc ();
+
+session_t *session_alloc_for_connection (transport_connection_t * tc);
+session_t *session_alloc_for_half_open (transport_connection_t *tc);
+
+typedef void (pool_safe_realloc_rpc_fn) (void *rpc_args);
+
+typedef struct
+{
+ u8 ph[STRUCT_OFFSET_OF (pool_header_t, max_elts) + 4];
+ u32 flag;
+} pool_safe_realloc_header_t;
+
+STATIC_ASSERT_SIZEOF (pool_safe_realloc_header_t, sizeof (pool_header_t));
+
+#define POOL_REALLOC_SAFE_ELT_THRESH 32
+
+#define pool_realloc_flag(PH) \
+ ((pool_safe_realloc_header_t *) pool_header (PH))->flag
+
+#define pool_realloc_safe_aligned(P, align) \
+ do \
+ { \
+ vlib_main_t *vm = vlib_get_main (); \
+ u32 free_elts, max_elts, n_alloc; \
+ ASSERT (vlib_get_thread_index () == 0); \
+ vlib_worker_thread_barrier_sync (vm); \
+ free_elts = pool_free_elts (P); \
+ max_elts = pool_max_len (P); \
+ n_alloc = clib_max (2 * max_elts, POOL_REALLOC_SAFE_ELT_THRESH); \
+ pool_alloc_aligned (P, free_elts + n_alloc, align); \
+ clib_bitmap_validate (pool_header (P)->free_bitmap, \
+ max_elts + n_alloc); \
+ pool_realloc_flag (P) = 0; \
+ vlib_worker_thread_barrier_release (vm); \
+ } \
+ while (0)
+
+always_inline void
+pool_program_safe_realloc (void *p, u32 thread_index,
+ pool_safe_realloc_rpc_fn *rpc_fn)
+{
+ /* Reuse pad as a realloc flag */
+ if (pool_realloc_flag (p))
+ return;
+
+ pool_realloc_flag (p) = 1;
+ session_send_rpc_evt_to_thread (0 /* thread index */, rpc_fn,
+ uword_to_pointer (thread_index, void *));
+}
+
+always_inline void
+pool_realloc_maybe_wait_at_barrier (void)
+{
+ if (!(*vlib_worker_threads->wait_at_barrier))
+ return;
+
+ /* Node refork required. Don't stop at the barrier from within a node */
+ if (*vlib_worker_threads->node_reforks_required)
+ return;
+
+ clib_atomic_fetch_add (vlib_worker_threads->workers_at_barrier, 1);
+
+ while (*vlib_worker_threads->wait_at_barrier)
+ ;
+
+ clib_atomic_fetch_add (vlib_worker_threads->workers_at_barrier, -1);
+}
+
+#define pool_realloc_all_at_barrier(_not) \
+ (*vlib_worker_threads->workers_at_barrier >= (vlib_num_workers () - _not))
+
+#define pool_realloc_safe_force(P) \
+ do \
+ { \
+ ALWAYS_ASSERT (*vlib_worker_threads->node_reforks_required); \
+ if (pool_realloc_all_at_barrier (1)) \
+ { \
+ pool_alloc (P, pool_max_len (P)); \
+ } \
+ else \
+ { \
+ session_main_t *sm = &session_main; \
+ clib_warning ("forced pool realloc"); \
+ clib_atomic_fetch_add (&sm->pool_realloc_at_barrier, 1); \
+ while (!pool_realloc_all_at_barrier (sm->pool_realloc_at_barrier)) \
+ ; \
+ clib_spinlock_lock (&sm->pool_realloc_lock); \
+ pool_alloc (P, pool_max_len (P)); \
+ clib_spinlock_unlock (&sm->pool_realloc_lock); \
+ clib_atomic_fetch_add (&sm->pool_realloc_at_barrier, -1); \
+ } \
+ } \
+ while (0)
+
+#define pool_needs_realloc(P) \
+ ((!P) || \
+ (vec_len (pool_header (P)->free_indices) < POOL_REALLOC_SAFE_ELT_THRESH && \
+ pool_free_elts (P) < POOL_REALLOC_SAFE_ELT_THRESH))
+
+#define pool_get_aligned_safe(P, E, thread_index, rpc_fn, align) \
+ do \
+ { \
+ ASSERT (vlib_get_thread_index () == thread_index || \
+ vlib_thread_is_main_w_barrier ()); \
+ if (PREDICT_FALSE (pool_needs_realloc (P))) \
+ { \
+ if (PREDICT_FALSE (!(P))) \
+ { \
+ pool_alloc_aligned (P, 2 * POOL_REALLOC_SAFE_ELT_THRESH, \
+ align); \
+ } \
+ else if (PREDICT_FALSE (pool_free_elts (P) < \
+ POOL_REALLOC_SAFE_ELT_THRESH / 2)) \
+ { \
+ volatile typeof (P) *PP = &(P); \
+ pool_program_safe_realloc (P, thread_index, rpc_fn); \
+ if (thread_index) \
+ { \
+ while (pool_realloc_flag (P)) \
+ { \
+ /* If refork required abort and consume existing elt */ \
+ if (*vlib_worker_threads->node_reforks_required) \
+ { \
+ /* All workers at barrier realloc now */ \
+ if (pool_realloc_all_at_barrier (1)) \
+ pool_alloc_aligned (P, pool_max_len (P), align); \
+ break; \
+ } \
+ pool_realloc_maybe_wait_at_barrier (); \
+ } \
+ if (pool_free_elts (P) == 0) \
+ pool_realloc_safe_force (P); \
+ ALWAYS_ASSERT (pool_free_elts (P) > 0); \
+ } \
+ (P) = *PP; \
+ } \
+ else \
+ { \
+ pool_program_safe_realloc (P, thread_index, rpc_fn); \
+ } \
+ } \
+ pool_get_aligned (P, E, align); \
+ } \
+ while (0)
#endif /* __included_session_h__ */