quic: update quicly to v0.1.3
[vpp.git] / src / plugins / quic / quic_crypto.c
index c30e68c..05211cd 100644 (file)
@@ -1,5 +1,5 @@
 /*
- * Copyright (c) 2019 Cisco and/or its affiliates.
+ * Copyright (c) 2021 Cisco and/or its affiliates.
  * Licensed under the Apache License, Version 2.0 (the "License");
  * you may not use this file except in compliance with the License.
  * You may obtain a copy of the License at:
  * limitations under the License.
  */
 
-#include <quic/quic_crypto.h>
 #include <quic/quic.h>
+#include <quic/quic_crypto.h>
 
-#include <vnet/crypto/crypto.h>
-
-#include <picotls/openssl.h>
 #include <quicly.h>
+#include <picotls/openssl.h>
 
-typedef void (*quicly_do_transform_fn) (ptls_cipher_context_t *, void *,
-                                       const void *, size_t);
+#define QUICLY_EPOCH_1RTT 3
+
+extern quic_main_t quic_main;
+extern quic_ctx_t *quic_get_conn_ctx (quicly_conn_t * conn);
+vnet_crypto_main_t *cm = &crypto_main;
 
 struct cipher_context_t
 {
   ptls_cipher_context_t super;
   vnet_crypto_op_t op;
+  vnet_crypto_op_id_t id;
   u32 key_index;
 };
 
@@ -36,195 +38,320 @@ struct aead_crypto_context_t
   ptls_aead_context_t super;
   vnet_crypto_op_t op;
   u32 key_index;
+  vnet_crypto_op_id_t id;
+  uint8_t iv[PTLS_MAX_IV_SIZE];
+  uint8_t static_iv[PTLS_MAX_IV_SIZE];
 };
 
-vnet_crypto_main_t *cm = &crypto_main;
-
-static void
-quic_crypto_cipher_do_init (ptls_cipher_context_t * _ctx, const void *iv)
+static int
+quic_crypto_setup_cipher (quicly_crypto_engine_t *engine, quicly_conn_t *conn,
+                         size_t epoch, int is_enc,
+                         ptls_cipher_context_t **header_protect_ctx,
+                         ptls_aead_context_t **packet_protect_ctx,
+                         ptls_aead_algorithm_t *aead,
+                         ptls_hash_algorithm_t *hash, const void *secret)
 {
-  struct cipher_context_t *ctx = (struct cipher_context_t *) _ctx;
+  uint8_t hpkey[PTLS_MAX_SECRET_SIZE];
+  int ret;
 
-  vnet_crypto_op_id_t id;
-  if (!strcmp (ctx->super.algo->name, "AES128-CTR"))
+  *packet_protect_ctx = NULL;
+  /* generate new header protection key */
+  if (header_protect_ctx != NULL)
     {
-      id = VNET_CRYPTO_OP_AES_128_CTR_ENC;
+      *header_protect_ctx = NULL;
+      ret =
+       ptls_hkdf_expand_label (hash, hpkey, aead->ctr_cipher->key_size,
+                               ptls_iovec_init (secret, hash->digest_size),
+                               "quic hp", ptls_iovec_init (NULL, 0), NULL);
+      if (ret)
+       goto Exit;
+      *header_protect_ctx = ptls_cipher_new (aead->ctr_cipher, is_enc, hpkey);
+      if (NULL == *header_protect_ctx)
+       {
+         ret = PTLS_ERROR_NO_MEMORY;
+         goto Exit;
+       }
     }
-  else if (!strcmp (ctx->super.algo->name, "AES256-CTR"))
+
+  /* generate new AEAD context */
+  *packet_protect_ctx =
+    ptls_aead_new (aead, hash, is_enc, secret, QUICLY_AEAD_BASE_LABEL);
+  if (NULL == *packet_protect_ctx)
     {
-      id = VNET_CRYPTO_OP_AES_256_CTR_ENC;
+      ret = PTLS_ERROR_NO_MEMORY;
+      goto Exit;
     }
-  else
+
+  if (epoch == QUICLY_EPOCH_1RTT && !is_enc)
     {
-      QUIC_DBG (1, "%s, Invalid crypto cipher : ", __FUNCTION__,
-               _ctx->algo->name);
-      assert (0);
+      quic_ctx_t *qctx = quic_get_conn_ctx (conn);
+      if (qctx->ingress_keys.aead_ctx != NULL)
+       qctx->key_phase_ingress++;
+
+      qctx->ingress_keys.aead_ctx = *packet_protect_ctx;
+      if (header_protect_ctx != NULL)
+       qctx->ingress_keys.hp_ctx = *header_protect_ctx;
     }
 
-  vnet_crypto_op_init (&ctx->op, id);
-  ctx->op.iv = (u8 *) iv;
-  ctx->op.key_index = ctx->key_index;
-}
+  ret = 0;
 
-static void
-quic_crypto_cipher_dispose (ptls_cipher_context_t * _ctx)
-{
-  /* Do nothing */
+Exit:
+  if (ret)
+    {
+      if (*packet_protect_ctx != NULL)
+       {
+         ptls_aead_free (*packet_protect_ctx);
+         *packet_protect_ctx = NULL;
+       }
+      if (header_protect_ctx && *header_protect_ctx != NULL)
+       {
+         ptls_cipher_free (*header_protect_ctx);
+         *header_protect_ctx = NULL;
+       }
+    }
+  ptls_clear_memory (hpkey, sizeof (hpkey));
+  return ret;
 }
 
-static void
-quic_crypto_cipher_encrypt (ptls_cipher_context_t * _ctx, void *output,
-                           const void *input, size_t _len)
+static size_t
+quic_crypto_aead_decrypt (quic_ctx_t *qctx, ptls_aead_context_t *_ctx,
+                         void *_output, const void *input, size_t inlen,
+                         uint64_t decrypted_pn, const void *aad,
+                         size_t aadlen)
 {
   vlib_main_t *vm = vlib_get_main ();
-  struct cipher_context_t *ctx = (struct cipher_context_t *) _ctx;
 
+  struct aead_crypto_context_t *ctx = (struct aead_crypto_context_t *) _ctx;
+
+  vnet_crypto_op_init (&ctx->op, ctx->id);
+  ctx->op.aad = (u8 *) aad;
+  ctx->op.aad_len = aadlen;
+  ctx->op.iv = ctx->iv;
+  ptls_aead__build_iv (ctx->super.algo, ctx->op.iv, ctx->static_iv,
+                      decrypted_pn);
   ctx->op.src = (u8 *) input;
-  ctx->op.dst = output;
-  ctx->op.len = _len;
+  ctx->op.dst = _output;
+  ctx->op.key_index = ctx->key_index;
+  ctx->op.len = inlen - ctx->super.algo->tag_size;
+  ctx->op.tag_len = ctx->super.algo->tag_size;
+  ctx->op.tag = ctx->op.src + ctx->op.len;
 
-  vnet_crypto_process_ops (vm, &ctx->op, 1);
+  vnet_crypto_process_ops (vm, &(ctx->op), 1);
+
+  return ctx->op.len;
 }
 
-static int
-quic_crypto_cipher_setup_crypto (ptls_cipher_context_t * _ctx, int is_enc,
-                                const void *key, const EVP_CIPHER * cipher,
-                                quicly_do_transform_fn do_transform)
+void
+quic_crypto_decrypt_packet (quic_ctx_t * qctx, quic_rx_packet_ctx_t * pctx)
 {
-  struct cipher_context_t *ctx = (struct cipher_context_t *) _ctx;
-
-  ctx->super.do_dispose = quic_crypto_cipher_dispose;
-  ctx->super.do_init = quic_crypto_cipher_do_init;
-  ctx->super.do_transform = do_transform;
-
-  vlib_main_t *vm = vlib_get_main ();
-  vnet_crypto_alg_t algo;
-  if (!strcmp (ctx->super.algo->name, "AES128-CTR"))
+  ptls_cipher_context_t *header_protection = NULL;
+  ptls_aead_context_t *aead = NULL;
+  int pn;
+
+  /* Long Header packets are not decrypted by vpp */
+  if (QUICLY_PACKET_IS_LONG_HEADER (pctx->packet.octets.base[0]))
+    return;
+
+  uint64_t next_expected_packet_number =
+    quicly_get_next_expected_packet_number (qctx->conn);
+  if (next_expected_packet_number == UINT64_MAX)
+    return;
+
+  aead = qctx->ingress_keys.aead_ctx;
+  header_protection = qctx->ingress_keys.hp_ctx;
+
+  if (!aead || !header_protection)
+    return;
+
+  size_t encrypted_len = pctx->packet.octets.len - pctx->packet.encrypted_off;
+  uint8_t hpmask[5] = { 0 };
+  uint32_t pnbits = 0;
+  size_t pnlen, ptlen, i;
+
+  /* decipher the header protection, as well as obtaining pnbits, pnlen */
+  if (encrypted_len < header_protection->algo->iv_size + QUICLY_MAX_PN_SIZE)
+    return;
+  ptls_cipher_init (header_protection,
+                   pctx->packet.octets.base + pctx->packet.encrypted_off +
+                   QUICLY_MAX_PN_SIZE);
+  ptls_cipher_encrypt (header_protection, hpmask, hpmask, sizeof (hpmask));
+  pctx->packet.octets.base[0] ^=
+    hpmask[0] & (QUICLY_PACKET_IS_LONG_HEADER (pctx->packet.octets.base[0]) ?
+                0xf : 0x1f);
+  pnlen = (pctx->packet.octets.base[0] & 0x3) + 1;
+  for (i = 0; i != pnlen; ++i)
     {
-      algo = VNET_CRYPTO_ALG_AES_128_CTR;
+      pctx->packet.octets.base[pctx->packet.encrypted_off + i] ^=
+       hpmask[i + 1];
+      pnbits =
+       (pnbits << 8) | pctx->packet.octets.base[pctx->packet.encrypted_off +
+                                                i];
     }
-  else if (!strcmp (ctx->super.algo->name, "AES256-CTR"))
+
+  size_t aead_off = pctx->packet.encrypted_off + pnlen;
+
+  pn =
+    quicly_determine_packet_number (pnbits, pnlen * 8,
+                                   next_expected_packet_number);
+
+  int key_phase_bit =
+    (pctx->packet.octets.base[0] & QUICLY_KEY_PHASE_BIT) != 0;
+
+  if (key_phase_bit != (qctx->key_phase_ingress & 1))
     {
-      algo = VNET_CRYPTO_ALG_AES_256_CTR;
+      pctx->packet.octets.base[0] ^=
+       hpmask[0] &
+       (QUICLY_PACKET_IS_LONG_HEADER (pctx->packet.octets.base[0]) ? 0xf :
+        0x1f);
+      for (i = 0; i != pnlen; ++i)
+       {
+         pctx->packet.octets.base[pctx->packet.encrypted_off + i] ^=
+           hpmask[i + 1];
+       }
+      return;
     }
-  else
+
+  if ((ptlen = quic_crypto_aead_decrypt (
+        qctx, aead, pctx->packet.octets.base + aead_off,
+        pctx->packet.octets.base + aead_off,
+        pctx->packet.octets.len - aead_off, pn, pctx->packet.octets.base,
+        aead_off)) == SIZE_MAX)
     {
-      QUIC_DBG (1, "%s, Invalid crypto cipher : ", __FUNCTION__,
-               _ctx->algo->name);
-      assert (0);
+      fprintf (stderr,
+              "%s: aead decryption failure (pn: %d)\n", __FUNCTION__, pn);
+      return;
     }
 
-  ctx->key_index = vnet_crypto_key_add (vm, algo,
-                                       (u8 *) key, _ctx->algo->key_size);
+  pctx->packet.encrypted_off = aead_off;
+  pctx->packet.octets.len = ptlen + aead_off;
 
-  return 0;
+  pctx->packet.decrypted.pn = pn;
+  pctx->packet.decrypted.key_phase = qctx->key_phase_ingress;
 }
 
-static int
-aes128ctr_setup_crypto (ptls_cipher_context_t * ctx, int is_enc,
-                       const void *key)
+void
+quic_crypto_encrypt_packet (struct st_quicly_crypto_engine_t *engine,
+                           quicly_conn_t *conn,
+                           ptls_cipher_context_t *header_protect_ctx,
+                           ptls_aead_context_t *packet_protect_ctx,
+                           ptls_iovec_t datagram, size_t first_byte_at,
+                           size_t payload_from, uint64_t packet_number,
+                           int coalesced)
 {
-  return quic_crypto_cipher_setup_crypto (ctx, 1, key, EVP_aes_128_ctr (),
-                                         quic_crypto_cipher_encrypt);
+  vlib_main_t *vm = vlib_get_main ();
+
+  struct cipher_context_t *hp_ctx =
+    (struct cipher_context_t *) header_protect_ctx;
+  struct aead_crypto_context_t *aead_ctx =
+    (struct aead_crypto_context_t *) packet_protect_ctx;
+
+  void *input = datagram.base + payload_from;
+  void *output = input;
+  size_t inlen =
+    datagram.len - payload_from - packet_protect_ctx->algo->tag_size;
+  const void *aad = datagram.base + first_byte_at;
+  size_t aadlen = payload_from - first_byte_at;
+
+  /* Build AEAD encrypt crypto operation */
+  vnet_crypto_op_init (&aead_ctx->op, aead_ctx->id);
+  aead_ctx->op.aad = (u8 *) aad;
+  aead_ctx->op.aad_len = aadlen;
+  aead_ctx->op.iv = aead_ctx->iv;
+  ptls_aead__build_iv (aead_ctx->super.algo, aead_ctx->op.iv,
+                      aead_ctx->static_iv, packet_number);
+  aead_ctx->op.key_index = aead_ctx->key_index;
+  aead_ctx->op.src = (u8 *) input;
+  aead_ctx->op.dst = output;
+  aead_ctx->op.len = inlen;
+  aead_ctx->op.tag_len = aead_ctx->super.algo->tag_size;
+  aead_ctx->op.tag = aead_ctx->op.src + inlen;
+  vnet_crypto_process_ops (vm, &(aead_ctx->op), 1);
+  assert (aead_ctx->op.status == VNET_CRYPTO_OP_STATUS_COMPLETED);
+
+  /* Build Header protection crypto operation */
+  ptls_aead_supplementary_encryption_t supp = {
+    .ctx = header_protect_ctx,
+    .input =
+      datagram.base + payload_from - QUICLY_SEND_PN_SIZE + QUICLY_MAX_PN_SIZE
+  };
+
+  /* Build Header protection crypto operation */
+  vnet_crypto_op_init (&hp_ctx->op, hp_ctx->id);
+  memset (supp.output, 0, sizeof (supp.output));
+  hp_ctx->op.iv = (u8 *) supp.input;
+  hp_ctx->op.key_index = hp_ctx->key_index;
+  hp_ctx->op.src = (u8 *) supp.output;
+  hp_ctx->op.dst = (u8 *) supp.output;
+  hp_ctx->op.len = sizeof (supp.output);
+  vnet_crypto_process_ops (vm, &(hp_ctx->op), 1);
+  assert (hp_ctx->op.status == VNET_CRYPTO_OP_STATUS_COMPLETED);
+
+  datagram.base[first_byte_at] ^=
+    supp.output[0] &
+    (QUICLY_PACKET_IS_LONG_HEADER (datagram.base[first_byte_at]) ? 0xf : 0x1f);
+  for (size_t i = 0; i != QUICLY_SEND_PN_SIZE; ++i)
+    datagram.base[payload_from + i - QUICLY_SEND_PN_SIZE] ^=
+      supp.output[i + 1];
 }
 
 static int
-aes256ctr_setup_crypto (ptls_cipher_context_t * ctx, int is_enc,
-                       const void *key)
+quic_crypto_cipher_setup_crypto (ptls_cipher_context_t *_ctx, int is_enc,
+                                const void *key, const EVP_CIPHER *cipher)
 {
-  return quic_crypto_cipher_setup_crypto (ctx, 1, key, EVP_aes_256_ctr (),
-                                         quic_crypto_cipher_encrypt);
-}
+  struct cipher_context_t *ctx = (struct cipher_context_t *) _ctx;
 
-size_t
-quic_crypto_aead_encrypt (ptls_aead_context_t * _ctx, void *output,
-                         const void *input, size_t inlen, uint64_t seq,
-                         const void *iv, const void *aad, size_t aadlen)
-{
   vlib_main_t *vm = vlib_get_main ();
-  struct aead_crypto_context_t *ctx = (struct aead_crypto_context_t *) _ctx;
-
-  vnet_crypto_op_id_t id;
-  if (!strcmp (ctx->super.algo->name, "AES128-GCM"))
+  vnet_crypto_alg_t algo;
+  if (!strcmp (ctx->super.algo->name, "AES128-CTR"))
     {
-      id = VNET_CRYPTO_OP_AES_128_GCM_ENC;
+      algo = VNET_CRYPTO_ALG_AES_128_CTR;
+      ctx->id = is_enc ? VNET_CRYPTO_OP_AES_128_CTR_ENC :
+                        VNET_CRYPTO_OP_AES_128_CTR_DEC;
+      ptls_openssl_aes128ctr.setup_crypto (_ctx, is_enc, key);
     }
-  else if (!strcmp (ctx->super.algo->name, "AES256-GCM"))
+  else if (!strcmp (ctx->super.algo->name, "AES256-CTR"))
     {
-      id = VNET_CRYPTO_OP_AES_256_GCM_ENC;
+      algo = VNET_CRYPTO_ALG_AES_256_CTR;
+      ctx->id = is_enc ? VNET_CRYPTO_OP_AES_256_CTR_ENC :
+                        VNET_CRYPTO_OP_AES_256_CTR_DEC;
+      ptls_openssl_aes256ctr.setup_crypto (_ctx, is_enc, key);
     }
   else
     {
+      QUIC_DBG (1, "%s, Invalid crypto cipher : ", __FUNCTION__,
+               _ctx->algo->name);
       assert (0);
     }
 
-  vnet_crypto_op_init (&ctx->op, id);
-  ctx->op.aad = (u8 *) aad;
-  ctx->op.aad_len = aadlen;
-  ctx->op.iv = (u8 *) iv;
-
-  ctx->op.src = (u8 *) input;
-  ctx->op.dst = output;
-  ctx->op.key_index = ctx->key_index;
-  ctx->op.len = inlen;
-
-  ctx->op.tag_len = ctx->super.algo->tag_size;
-  ctx->op.tag = ctx->op.src + inlen;
-
-  vnet_crypto_process_ops (vm, &ctx->op, 1);
-
-  return ctx->op.len + ctx->op.tag_len;
-}
-
-size_t
-quic_crypto_aead_decrypt (ptls_aead_context_t * _ctx, void *_output,
-                         const void *input, size_t inlen, const void *iv,
-                         const void *aad, size_t aadlen)
-{
-  vlib_main_t *vm = vlib_get_main ();
-  struct aead_crypto_context_t *ctx = (struct aead_crypto_context_t *) _ctx;
-
-  vnet_crypto_op_id_t id;
-  if (!strcmp (ctx->super.algo->name, "AES128-GCM"))
+  if (quic_main.vnet_crypto_enabled)
     {
-      id = VNET_CRYPTO_OP_AES_128_GCM_DEC;
-    }
-  else if (!strcmp (ctx->super.algo->name, "AES256-GCM"))
-    {
-      id = VNET_CRYPTO_OP_AES_256_GCM_DEC;
-    }
-  else
-    {
-      assert (0);
+      clib_rwlock_writer_lock (&quic_main.crypto_keys_quic_rw_lock);
+      ctx->key_index =
+       vnet_crypto_key_add (vm, algo, (u8 *) key, _ctx->algo->key_size);
+      clib_rwlock_writer_unlock (&quic_main.crypto_keys_quic_rw_lock);
     }
 
-  vnet_crypto_op_init (&ctx->op, id);
-  ctx->op.aad = (u8 *) aad;
-  ctx->op.aad_len = aadlen;
-  ctx->op.iv = (u8 *) iv;
-
-  ctx->op.src = (u8 *) input;
-  ctx->op.dst = _output;
-  ctx->op.key_index = ctx->key_index;
-  ctx->op.len = inlen - ctx->super.algo->tag_size;
-
-  ctx->op.tag_len = ctx->super.algo->tag_size;
-  ctx->op.tag = ctx->op.src + ctx->op.len;
-
-  vnet_crypto_process_ops (vm, &ctx->op, 1);
-
-  return ctx->op.len;
+  return 0;
 }
 
-static void
-quic_crypto_aead_dispose_crypto (ptls_aead_context_t * _ctx)
+static int
+quic_crypto_aes128ctr_setup_crypto (ptls_cipher_context_t * ctx, int is_enc,
+                                   const void *key)
 {
+  return quic_crypto_cipher_setup_crypto (ctx, 1, key, EVP_aes_128_ctr ());
+}
 
+static int
+quic_crypto_aes256ctr_setup_crypto (ptls_cipher_context_t * ctx, int is_enc,
+                                   const void *key)
+{
+  return quic_crypto_cipher_setup_crypto (ctx, 1, key, EVP_aes_256_ctr ());
 }
 
 static int
-quic_crypto_aead_setup_crypto (ptls_aead_context_t * _ctx, int is_enc,
-                              const void *key, const EVP_CIPHER * cipher)
+quic_crypto_aead_setup_crypto (ptls_aead_context_t *_ctx, int is_enc,
+                              const void *key, const void *iv,
+                              const EVP_CIPHER *cipher)
 {
   vlib_main_t *vm = vlib_get_main ();
   struct aead_crypto_context_t *ctx = (struct aead_crypto_context_t *) _ctx;
@@ -233,10 +360,16 @@ quic_crypto_aead_setup_crypto (ptls_aead_context_t * _ctx, int is_enc,
   if (!strcmp (ctx->super.algo->name, "AES128-GCM"))
     {
       algo = VNET_CRYPTO_ALG_AES_128_GCM;
+      ctx->id = is_enc ? VNET_CRYPTO_OP_AES_128_GCM_ENC :
+                        VNET_CRYPTO_OP_AES_128_GCM_DEC;
+      ptls_openssl_aes128gcm.setup_crypto (_ctx, is_enc, key, iv);
     }
   else if (!strcmp (ctx->super.algo->name, "AES256-GCM"))
     {
       algo = VNET_CRYPTO_ALG_AES_256_GCM;
+      ctx->id = is_enc ? VNET_CRYPTO_OP_AES_256_GCM_ENC :
+                        VNET_CRYPTO_OP_AES_256_GCM_DEC;
+      ptls_openssl_aes256gcm.setup_crypto (_ctx, is_enc, key, iv);
     }
   else
     {
@@ -245,48 +378,102 @@ quic_crypto_aead_setup_crypto (ptls_aead_context_t * _ctx, int is_enc,
       assert (0);
     }
 
-  ctx->super.do_decrypt = quic_crypto_aead_decrypt;
-  ctx->super.do_encrypt = quic_crypto_aead_encrypt;
-  ctx->super.dispose_crypto = quic_crypto_aead_dispose_crypto;
+  if (quic_main.vnet_crypto_enabled)
+    {
+      clib_memcpy (ctx->static_iv, iv, ctx->super.algo->iv_size);
 
-  ctx->key_index = vnet_crypto_key_add (vm, algo,
-                                       (u8 *) key, _ctx->algo->key_size);
+      clib_rwlock_writer_lock (&quic_main.crypto_keys_quic_rw_lock);
+      ctx->key_index = vnet_crypto_key_add (vm, algo,
+                                           (u8 *) key, _ctx->algo->key_size);
+      clib_rwlock_writer_unlock (&quic_main.crypto_keys_quic_rw_lock);
+    }
 
   return 0;
 }
 
 static int
-quic_crypto_aead_aes128gcm_setup_crypto (ptls_aead_context_t * ctx,
-                                        int is_enc, const void *key)
+quic_crypto_aead_aes128gcm_setup_crypto (ptls_aead_context_t *ctx, int is_enc,
+                                        const void *key, const void *iv)
 {
-  return quic_crypto_aead_setup_crypto (ctx, is_enc, key, EVP_aes_128_gcm ());
+  return quic_crypto_aead_setup_crypto (ctx, is_enc, key, iv,
+                                       EVP_aes_128_gcm ());
 }
 
 static int
-quic_crypto_aead_aes256gcm_setup_crypto (ptls_aead_context_t * ctx,
-                                        int is_enc, const void *key)
+quic_crypto_aead_aes256gcm_setup_crypto (ptls_aead_context_t *ctx, int is_enc,
+                                        const void *key, const void *iv)
 {
-  return quic_crypto_aead_setup_crypto (ctx, is_enc, key, EVP_aes_256_gcm ());
+  return quic_crypto_aead_setup_crypto (ctx, is_enc, key, iv,
+                                       EVP_aes_256_gcm ());
 }
 
-ptls_cipher_algorithm_t quic_crypto_aes128ctr = { "AES128-CTR",
+int
+quic_encrypt_ticket_cb (ptls_encrypt_ticket_t *_self, ptls_t *tls,
+                       int is_encrypt, ptls_buffer_t *dst, ptls_iovec_t src)
+{
+  quic_session_cache_t *self = (void *) _self;
+  int ret;
+
+  if (is_encrypt)
+    {
+
+      /* replace the cached entry along with a newly generated session id */
+      clib_mem_free (self->data.base);
+      if ((self->data.base = clib_mem_alloc (src.len)) == NULL)
+       return PTLS_ERROR_NO_MEMORY;
+
+      ptls_get_context (tls)->random_bytes (self->id, sizeof (self->id));
+      clib_memcpy (self->data.base, src.base, src.len);
+      self->data.len = src.len;
+
+      /* store the session id in buffer */
+      if ((ret = ptls_buffer_reserve (dst, sizeof (self->id))) != 0)
+       return ret;
+      clib_memcpy (dst->base + dst->off, self->id, sizeof (self->id));
+      dst->off += sizeof (self->id);
+    }
+  else
+    {
+      /* check if session id is the one stored in cache */
+      if (src.len != sizeof (self->id))
+       return PTLS_ERROR_SESSION_NOT_FOUND;
+      if (clib_memcmp (self->id, src.base, sizeof (self->id)) != 0)
+       return PTLS_ERROR_SESSION_NOT_FOUND;
+
+      /* return the cached value */
+      if ((ret = ptls_buffer_reserve (dst, self->data.len)) != 0)
+       return ret;
+      clib_memcpy (dst->base + dst->off, self->data.base, self->data.len);
+      dst->off += self->data.len;
+    }
+
+  return 0;
+}
+
+ptls_cipher_algorithm_t quic_crypto_aes128ctr = {
+  "AES128-CTR",
   PTLS_AES128_KEY_SIZE,
-  1, PTLS_AES_IV_SIZE,
+  1,
+  PTLS_AES_IV_SIZE,
   sizeof (struct cipher_context_t),
-  aes128ctr_setup_crypto
+  quic_crypto_aes128ctr_setup_crypto
 };
 
-ptls_cipher_algorithm_t quic_crypto_aes256ctr = { "AES256-CTR",
+ptls_cipher_algorithm_t quic_crypto_aes256ctr = {
+  "AES256-CTR",
   PTLS_AES256_KEY_SIZE,
-  1 /* block size */ ,
+  1 /* block size */,
   PTLS_AES_IV_SIZE,
   sizeof (struct cipher_context_t),
-  aes256ctr_setup_crypto
+  quic_crypto_aes256ctr_setup_crypto
 };
 
-ptls_aead_algorithm_t quic_crypto_aes128gcm = { "AES128-GCM",
+ptls_aead_algorithm_t quic_crypto_aes128gcm = {
+  "AES128-GCM",
+  PTLS_AESGCM_CONFIDENTIALITY_LIMIT,
+  PTLS_AESGCM_INTEGRITY_LIMIT,
   &quic_crypto_aes128ctr,
-  NULL,
+  &ptls_openssl_aes128ecb,
   PTLS_AES128_KEY_SIZE,
   PTLS_AESGCM_IV_SIZE,
   PTLS_AESGCM_TAG_SIZE,
@@ -294,9 +481,12 @@ ptls_aead_algorithm_t quic_crypto_aes128gcm = { "AES128-GCM",
   quic_crypto_aead_aes128gcm_setup_crypto
 };
 
-ptls_aead_algorithm_t quic_crypto_aes256gcm = { "AES256-GCM",
+ptls_aead_algorithm_t quic_crypto_aes256gcm = {
+  "AES256-GCM",
+  PTLS_AESGCM_CONFIDENTIALITY_LIMIT,
+  PTLS_AESGCM_INTEGRITY_LIMIT,
   &quic_crypto_aes256ctr,
-  NULL,
+  &ptls_openssl_aes256ecb,
   PTLS_AES256_KEY_SIZE,
   PTLS_AESGCM_IV_SIZE,
   PTLS_AESGCM_TAG_SIZE,
@@ -304,24 +494,23 @@ ptls_aead_algorithm_t quic_crypto_aes256gcm = { "AES256-GCM",
   quic_crypto_aead_aes256gcm_setup_crypto
 };
 
-ptls_cipher_suite_t quic_crypto_aes128gcmsha256 =
-  { PTLS_CIPHER_SUITE_AES_128_GCM_SHA256,
-  &quic_crypto_aes128gcm,
-  &ptls_openssl_sha256
+ptls_cipher_suite_t quic_crypto_aes128gcmsha256 = {
+  PTLS_CIPHER_SUITE_AES_128_GCM_SHA256,
+  &quic_crypto_aes128gcm, &ptls_openssl_sha256
 };
 
-ptls_cipher_suite_t quic_crypto_aes256gcmsha384 =
-  { PTLS_CIPHER_SUITE_AES_256_GCM_SHA384,
-  &quic_crypto_aes256gcm,
-  &ptls_openssl_sha384
+ptls_cipher_suite_t quic_crypto_aes256gcmsha384 = {
+  PTLS_CIPHER_SUITE_AES_256_GCM_SHA384,
+  &quic_crypto_aes256gcm, &ptls_openssl_sha384
 };
 
-ptls_cipher_suite_t *quic_crypto_cipher_suites[] =
-  { &quic_crypto_aes256gcmsha384,
-  &quic_crypto_aes128gcmsha256,
-  NULL
+ptls_cipher_suite_t *quic_crypto_cipher_suites[] = {
+  &quic_crypto_aes256gcmsha384, &quic_crypto_aes128gcmsha256, NULL
 };
 
+quicly_crypto_engine_t quic_crypto_engine = { quic_crypto_setup_cipher,
+                                             quic_crypto_encrypt_packet };
+
 /*
  * fd.io coding-style-patch-verification: ON
  *