#include <vnet/tls/tls.h>
#include <ctype.h>
#include <tlsopenssl/tls_openssl.h>
-#include <tlsopenssl/tls_bio.h>
+#include <tlsopenssl/tls_bios.h>
#define MAX_CRYPTO_LEN 64
openssl_main_t openssl_main;
+
static u32
-openssl_ctx_alloc (void)
+openssl_ctx_alloc_w_thread (u32 thread_index)
{
- u8 thread_index = vlib_get_thread_index ();
- openssl_main_t *tm = &openssl_main;
+ openssl_main_t *om = &openssl_main;
openssl_ctx_t **ctx;
- pool_get (tm->ctx_pool[thread_index], ctx);
+ pool_get (om->ctx_pool[thread_index], ctx);
if (!(*ctx))
*ctx = clib_mem_alloc (sizeof (openssl_ctx_t));
(*ctx)->ctx.c_thread_index = thread_index;
(*ctx)->ctx.tls_ctx_engine = CRYPTO_ENGINE_OPENSSL;
(*ctx)->ctx.app_session_handle = SESSION_INVALID_HANDLE;
- (*ctx)->openssl_ctx_index = ctx - tm->ctx_pool[thread_index];
+ (*ctx)->openssl_ctx_index = ctx - om->ctx_pool[thread_index];
return ((*ctx)->openssl_ctx_index);
}
+static u32
+openssl_ctx_alloc (void)
+{
+ return openssl_ctx_alloc_w_thread (vlib_get_thread_index ());
+}
+
static void
openssl_ctx_free (tls_ctx_t * ctx)
{
openssl_ctx_t *oc = (openssl_ctx_t *) ctx;
- if (SSL_is_init_finished (oc->ssl) && !ctx->is_passive_close)
- SSL_shutdown (oc->ssl);
+ /* Cleanup ssl ctx unless migrated */
+ if (!ctx->is_migrated)
+ {
+ if (SSL_is_init_finished (oc->ssl) && !ctx->is_passive_close)
+ SSL_shutdown (oc->ssl);
- SSL_free (oc->ssl);
+ SSL_free (oc->ssl);
+ vec_free (ctx->srv_hostname);
#ifdef HAVE_OPENSSL_ASYNC
openssl_evt_free (ctx->evt_index, ctx->c_thread_index);
#endif
- vec_free (ctx->srv_hostname);
+ }
+
pool_put_index (openssl_main.ctx_pool[ctx->c_thread_index],
oc->openssl_ctx_index);
}
+static void *
+openssl_ctx_detach (tls_ctx_t *ctx)
+{
+ openssl_ctx_t *oc = (openssl_ctx_t *) ctx, *oc_copy;
+
+ oc_copy = clib_mem_alloc (sizeof (*oc_copy));
+ clib_memcpy (oc_copy, oc, sizeof (*oc));
+
+ return oc_copy;
+}
+
+static u32
+openssl_ctx_attach (u32 thread_index, void *ctx_ptr)
+{
+ openssl_main_t *om = &openssl_main;
+ session_handle_t sh;
+ openssl_ctx_t **oc;
+
+ pool_get (om->ctx_pool[thread_index], oc);
+ /* Free the old instance instead of looking for an empty spot */
+ if (*oc)
+ clib_mem_free (*oc);
+
+ *oc = ctx_ptr;
+ (*oc)->openssl_ctx_index = oc - om->ctx_pool[thread_index];
+ (*oc)->ctx.c_thread_index = thread_index;
+
+ sh = (*oc)->ctx.tls_session_handle;
+ BIO_set_data ((*oc)->rbio, uword_to_pointer (sh, void *));
+ BIO_set_data ((*oc)->wbio, uword_to_pointer (sh, void *));
+
+ return ((*oc)->openssl_ctx_index);
+}
+
tls_ctx_t *
openssl_ctx_get (u32 ctx_index)
{
u32 max_enq;
max_enq = svm_fifo_max_enqueue_prod (f);
+ if (!max_enq)
+ return 0;
+
n_fs = svm_fifo_provision_chunks (f, fs, n_segs, max_enq);
if (n_fs < 0)
return 0;
rv = SSL_read (ssl, fs[i].data, fs[i].len);
read += rv > 0 ? rv : 0;
- if (rv < fs[i].len)
+ if (rv < (int) fs[i].len)
break;
}
{
rv = SSL_write (ssl, fs[i].data, fs[i].len);
wrote += (rv > 0) ? rv : 0;
- if (rv < fs[i].len)
+ if (rv < (int) fs[i].len)
break;
i++;
}
{
/* Need to check transport status */
if (ctx->is_passive_close)
- openssl_handle_handshake_failure (ctx);
- else
- tls_notify_app_accept (ctx);
+ {
+ openssl_handle_handshake_failure (ctx);
+ return -1;
+ }
+
+ /* Accept failed, cleanup */
+ if (tls_notify_app_accept (ctx))
+ {
+ ctx->c_s_index = SESSION_INVALID_INDEX;
+ tls_disconnect_transport (ctx);
+ return -1;
+ }
}
TLS_DBG (1, "Handshake for %u complete. TLS cipher is %s",
session_transport_closed_notify (&ctx->connection);
}
-static inline int
-openssl_ctx_write (tls_ctx_t * ctx, session_t * app_session,
- transport_send_params_t * sp)
+static int
+openssl_ctx_write_tls (tls_ctx_t *ctx, session_t *app_session,
+ transport_send_params_t *sp)
{
openssl_ctx_t *oc = (openssl_ctx_t *) ctx;
u32 deq_max, space, enq_buf;
deq_max = clib_min (deq_max, sp->max_burst_size);
+ /* Make sure tcp's tx fifo can actually buffer all bytes to be dequeued.
+ * If under memory pressure, tls's fifo segment might not be able to
+ * allocate the chunks needed. This also avoids errors from the underlying
+ * custom bio to the ssl infra which at times can get stuck. */
+ if (svm_fifo_provision_chunks (ts->tx_fifo, 0, 0, deq_max + TLSO_CTRL_BYTES))
+ goto check_tls_fifo;
+
wrote = openssl_write_from_fifo_into_ssl (f, oc->ssl, deq_max);
if (!wrote)
goto check_tls_fifo;
check_tls_fifo:
+ if (PREDICT_FALSE (ctx->app_closed && BIO_ctrl_pending (oc->rbio) <= 0))
+ openssl_confirm_app_close (ctx);
+
/* Deschedule and wait for deq notification if fifo is almost full */
enq_buf = clib_min (svm_fifo_size (ts->tx_fifo) / 2, TLSO_MIN_ENQ_SPACE);
if (space < wrote + enq_buf)
return wrote;
}
+static int
+openssl_ctx_write_dtls (tls_ctx_t *ctx, session_t *app_session,
+ transport_send_params_t *sp)
+{
+ openssl_main_t *om = &openssl_main;
+ openssl_ctx_t *oc = (openssl_ctx_t *) ctx;
+ u32 read = 0, to_deq, dgram_sz, enq_max;
+ session_dgram_pre_hdr_t hdr;
+ session_t *us;
+ int wrote, rv;
+ u8 *buf;
+
+ us = session_get_from_handle (ctx->tls_session_handle);
+ to_deq = svm_fifo_max_dequeue_cons (app_session->tx_fifo);
+ buf = om->tx_bufs[ctx->c_thread_index];
+
+ while (to_deq > 0)
+ {
+ /* Peeking only pre-header dgram because the session is connected */
+ rv = svm_fifo_peek (app_session->tx_fifo, 0, sizeof (hdr), (u8 *) &hdr);
+ ASSERT (rv == sizeof (hdr) && hdr.data_length < vec_len (buf));
+ ASSERT (to_deq >= hdr.data_length + SESSION_CONN_HDR_LEN);
+
+ dgram_sz = hdr.data_length + SESSION_CONN_HDR_LEN;
+ enq_max = dgram_sz + TLSO_CTRL_BYTES;
+ if (svm_fifo_max_enqueue_prod (us->tx_fifo) < enq_max ||
+ svm_fifo_provision_chunks (us->tx_fifo, 0, 0, enq_max))
+ {
+ svm_fifo_add_want_deq_ntf (us->tx_fifo, SVM_FIFO_WANT_DEQ_NOTIF);
+ transport_connection_deschedule (&ctx->connection);
+ sp->flags |= TRANSPORT_SND_F_DESCHED;
+ goto done;
+ }
+
+ rv = svm_fifo_peek (app_session->tx_fifo, SESSION_CONN_HDR_LEN,
+ hdr.data_length, buf);
+ ASSERT (rv == hdr.data_length);
+ svm_fifo_dequeue_drop (app_session->tx_fifo, dgram_sz);
+
+ wrote = SSL_write (oc->ssl, buf, rv);
+ ASSERT (wrote > 0);
+
+ read += rv;
+ to_deq -= dgram_sz;
+ }
+
+done:
+
+ if (svm_fifo_needs_deq_ntf (app_session->tx_fifo, read))
+ session_dequeue_notify (app_session);
+
+ if (read)
+ tls_add_vpp_q_tx_evt (us);
+
+ if (PREDICT_FALSE (ctx->app_closed &&
+ !svm_fifo_max_enqueue_prod (us->rx_fifo)))
+ openssl_confirm_app_close (ctx);
+
+ return read;
+}
+
static inline int
-openssl_ctx_read (tls_ctx_t * ctx, session_t * tls_session)
+openssl_ctx_write (tls_ctx_t *ctx, session_t *app_session,
+ transport_send_params_t *sp)
+{
+ if (ctx->tls_type == TRANSPORT_PROTO_TLS)
+ return openssl_ctx_write_tls (ctx, app_session, sp);
+ else
+ return openssl_ctx_write_dtls (ctx, app_session, sp);
+}
+
+static inline int
+openssl_ctx_read_tls (tls_ctx_t *ctx, session_t *tls_session)
{
openssl_ctx_t *oc = (openssl_ctx_t *) ctx;
session_t *app_session;
- int read, wrote = 0;
+ int read;
svm_fifo_t *f;
if (PREDICT_FALSE (SSL_in_init (oc->ssl)))
if (read && app_session->session_state >= SESSION_STATE_READY)
tls_notify_app_enqueue (ctx, app_session);
- if (SSL_pending (oc->ssl) > 0)
+ if ((SSL_pending (oc->ssl) > 0) ||
+ svm_fifo_max_dequeue_cons (tls_session->rx_fifo))
tls_add_vpp_q_builtin_rx_evt (tls_session);
+ return read;
+}
+
+static inline int
+openssl_ctx_read_dtls (tls_ctx_t *ctx, session_t *us)
+{
+ openssl_main_t *om = &openssl_main;
+ openssl_ctx_t *oc = (openssl_ctx_t *) ctx;
+ session_dgram_hdr_t hdr;
+ session_t *app_session;
+ u32 wrote = 0;
+ int read, rv;
+ u8 *buf;
+
+ if (PREDICT_FALSE (SSL_in_init (oc->ssl)))
+ {
+ u32 us_index = us->session_index;
+ if (openssl_ctx_handshake_rx (ctx, us) < 0)
+ return 0;
+ /* Session pool might grow when allocating the app's session */
+ us = session_get (us_index, ctx->c_thread_index);
+ }
+
+ buf = om->rx_bufs[ctx->c_thread_index];
+ app_session = session_get_from_handle (ctx->app_session_handle);
+ svm_fifo_fill_chunk_list (app_session->rx_fifo);
+
+ while (svm_fifo_max_dequeue_cons (us->rx_fifo) > 0)
+ {
+ if (svm_fifo_max_enqueue_prod (app_session->rx_fifo) < DTLSO_MAX_DGRAM)
+ {
+ tls_add_vpp_q_builtin_rx_evt (us);
+ goto done;
+ }
+
+ read = SSL_read (oc->ssl, buf, vec_len (buf));
+ if (PREDICT_FALSE (read <= 0))
+ {
+ if (read < 0)
+ tls_add_vpp_q_builtin_rx_evt (us);
+ goto done;
+ }
+ wrote += read;
+
+ hdr.data_length = read;
+ hdr.data_offset = 0;
+
+ svm_fifo_seg_t segs[2] = { { (u8 *) &hdr, sizeof (hdr) },
+ { buf, read } };
+
+ rv = svm_fifo_enqueue_segments (app_session->rx_fifo, segs, 2,
+ 0 /* allow partial */);
+ ASSERT (rv > 0);
+ }
+
+done:
+
+ /* If handshake just completed, session may still be in accepting state */
+ if (app_session->session_state >= SESSION_STATE_READY)
+ tls_notify_app_enqueue (ctx, app_session);
+
return wrote;
}
+static inline int
+openssl_ctx_read (tls_ctx_t *ctx, session_t *ts)
+{
+ if (ctx->tls_type == TRANSPORT_PROTO_TLS)
+ return openssl_ctx_read_tls (ctx, ts);
+ else
+ return openssl_ctx_read_dtls (ctx, ts);
+}
+
static int
openssl_ctx_init_client (tls_ctx_t * ctx)
{
const SSL_METHOD *method;
int rv, err;
- method = SSLv23_client_method ();
+ method = ctx->tls_type == TRANSPORT_PROTO_TLS ? SSLv23_client_method () :
+ DTLS_client_method ();
if (method == NULL)
{
- TLS_DBG (1, "SSLv23_method returned null");
+ TLS_DBG (1, "(D)TLS_method returned null");
return -1;
}
return -1;
}
- oc->rbio = BIO_new_tls (ctx->tls_session_handle);
- oc->wbio = BIO_new_tls (ctx->tls_session_handle);
+ if (ctx->tls_type == TRANSPORT_PROTO_TLS)
+ {
+ oc->rbio = BIO_new_tls (ctx->tls_session_handle);
+ oc->wbio = BIO_new_tls (ctx->tls_session_handle);
+ }
+ else
+ {
+ oc->rbio = BIO_new_dtls (ctx->tls_session_handle);
+ oc->wbio = BIO_new_dtls (ctx->tls_session_handle);
+ }
SSL_set_bio (oc->ssl, oc->wbio, oc->rbio);
SSL_set_connect_state (oc->ssl);
return -1;
}
- method = SSLv23_method ();
+ method = lctx->tls_type == TRANSPORT_PROTO_TLS ? SSLv23_server_method () :
+ DTLS_server_method ();
ssl_ctx = SSL_CTX_new (method);
if (!ssl_ctx)
{
* Set the key and cert
*/
cert_bio = BIO_new (BIO_s_mem ());
+ if (!cert_bio)
+ {
+ clib_warning ("unable to allocate memory");
+ return -1;
+ }
BIO_write (cert_bio, ckpair->cert, vec_len (ckpair->cert));
srvcert = PEM_read_bio_X509 (cert_bio, NULL, NULL, NULL);
if (!srvcert)
{
clib_warning ("unable to parse certificate");
- return -1;
+ goto err;
+ }
+ rv = SSL_CTX_use_certificate (ssl_ctx, srvcert);
+ if (rv != 1)
+ {
+ clib_warning ("unable to use SSL certificate");
+ goto err;
}
- SSL_CTX_use_certificate (ssl_ctx, srvcert);
+
BIO_free (cert_bio);
cert_bio = BIO_new (BIO_s_mem ());
+ if (!cert_bio)
+ {
+ clib_warning ("unable to allocate memory");
+ return -1;
+ }
BIO_write (cert_bio, ckpair->key, vec_len (ckpair->key));
pkey = PEM_read_bio_PrivateKey (cert_bio, NULL, NULL, NULL);
if (!pkey)
{
clib_warning ("unable to parse pkey");
- return -1;
+ goto err;
}
- SSL_CTX_use_PrivateKey (ssl_ctx, pkey);
+ rv = SSL_CTX_use_PrivateKey (ssl_ctx, pkey);
+ if (rv != 1)
+ {
+ clib_warning ("unable to use SSL PrivateKey");
+ goto err;
+ }
+
BIO_free (cert_bio);
olc_index = openssl_listen_ctx_alloc ();
return 0;
+err:
+ if (cert_bio)
+ BIO_free (cert_bio);
+ return -1;
}
static int
return -1;
}
- oc->rbio = BIO_new_tls (ctx->tls_session_handle);
- oc->wbio = BIO_new_tls (ctx->tls_session_handle);
+ if (ctx->tls_type == TRANSPORT_PROTO_TLS)
+ {
+ oc->rbio = BIO_new_tls (ctx->tls_session_handle);
+ oc->wbio = BIO_new_tls (ctx->tls_session_handle);
+ }
+ else
+ {
+ oc->rbio = BIO_new_dtls (ctx->tls_session_handle);
+ oc->wbio = BIO_new_dtls (ctx->tls_session_handle);
+ }
SSL_set_bio (oc->ssl, oc->wbio, oc->rbio);
SSL_set_accept_state (oc->ssl);
const static tls_engine_vft_t openssl_engine = {
.ctx_alloc = openssl_ctx_alloc,
+ .ctx_alloc_w_thread = openssl_ctx_alloc_w_thread,
.ctx_free = openssl_ctx_free,
+ .ctx_attach = openssl_ctx_attach,
+ .ctx_detach = openssl_ctx_detach,
.ctx_get = openssl_ctx_get,
.ctx_get_w_thread = openssl_ctx_get_w_thread,
.ctx_init_server = openssl_ctx_init_server,
vlib_thread_main_t *vtm = vlib_get_thread_main ();
openssl_main_t *om = &openssl_main;
clib_error_t *error = 0;
- u32 num_threads;
+ u32 num_threads, i;
error = tls_openssl_api_init (vm);
num_threads = 1 /* main thread */ + vtm->n_threads;
}
vec_validate (om->ctx_pool, num_threads - 1);
-
+ vec_validate (om->rx_bufs, num_threads - 1);
+ vec_validate (om->tx_bufs, num_threads - 1);
+ for (i = 0; i < num_threads; i++)
+ {
+ vec_validate (om->rx_bufs[i], DTLSO_MAX_DGRAM);
+ vec_validate (om->tx_bufs[i], DTLSO_MAX_DGRAM);
+ }
tls_register_engine (&openssl_engine, CRYPTO_ENGINE_OPENSSL);
om->engine_init = 0;