X-Git-Url: https://gerrit.fd.io/r/gitweb?a=blobdiff_plain;f=src%2Fvnet%2Fipsec%2Fesp.h;h=b1736453b93f97fdc9c293b380376cd8a5c95707;hb=14b804898a5e5b9ec5029d3466621178b85af925;hp=17300383e16a3e4654fb66d43bc6875ba68aa839;hpb=91f17dc7c4e79343b8fba924e7cea6380a0e6653;p=vpp.git diff --git a/src/vnet/ipsec/esp.h b/src/vnet/ipsec/esp.h index 17300383e16..b1736453b93 100644 --- a/src/vnet/ipsec/esp.h +++ b/src/vnet/ipsec/esp.h @@ -21,7 +21,11 @@ typedef struct { - u32 spi; + union + { + u32 spi; + u8 spi_bytes[4]; + }; u32 seq; u8 data[0]; } esp_header_t; @@ -54,140 +58,37 @@ typedef CLIB_PACKED (struct { }) ip6_and_esp_header_t; /* *INDENT-ON* */ -#define ESP_WINDOW_SIZE (64) -#define ESP_SEQ_MAX (4294967295UL) - -u8 *format_esp_header (u8 * s, va_list * args); - -always_inline int -esp_replay_check (ipsec_sa_t * sa, u32 seq) -{ - u32 diff; - - if (PREDICT_TRUE (seq > sa->last_seq)) - return 0; - - diff = sa->last_seq - seq; - - if (ESP_WINDOW_SIZE > diff) - return (sa->replay_window & (1ULL << diff)) ? 1 : 0; - else - return 1; - - return 0; -} - -always_inline int -esp_replay_check_esn (ipsec_sa_t * sa, u32 seq) +/** + * AES GCM Additional Authentication data + */ +typedef struct esp_aead_t_ { - u32 tl = sa->last_seq; - u32 th = sa->last_seq_hi; - u32 diff = tl - seq; - - if (PREDICT_TRUE (tl >= (ESP_WINDOW_SIZE - 1))) - { - if (seq >= (tl - ESP_WINDOW_SIZE + 1)) - { - sa->seq_hi = th; - if (seq <= tl) - return (sa->replay_window & (1ULL << diff)) ? 1 : 0; - else - return 0; - } - else - { - sa->seq_hi = th + 1; - return 0; - } - } - else - { - if (seq >= (tl - ESP_WINDOW_SIZE + 1)) - { - sa->seq_hi = th - 1; - return (sa->replay_window & (1ULL << diff)) ? 1 : 0; - } - else - { - sa->seq_hi = th; - if (seq <= tl) - return (sa->replay_window & (1ULL << diff)) ? 1 : 0; - else - return 0; - } - } + /** + * for GCM: when using ESN it's: + * SPI, seq-hi, seg-low + * else + * SPI, seq-low + */ + u32 data[3]; +} __clib_packed esp_aead_t; + +#define ESP_SEQ_MAX (4294967295UL) +#define ESP_MAX_BLOCK_SIZE (16) +#define ESP_MAX_IV_SIZE (16) +#define ESP_MAX_ICV_SIZE (32) - return 0; -} +u8 *format_esp_header (u8 * s, va_list * args); /* TODO seq increment should be atomic to be accessed by multiple workers */ -always_inline void -esp_replay_advance (ipsec_sa_t * sa, u32 seq) -{ - u32 pos; - - if (seq > sa->last_seq) - { - pos = seq - sa->last_seq; - if (pos < ESP_WINDOW_SIZE) - sa->replay_window = ((sa->replay_window) << pos) | 1; - else - sa->replay_window = 1; - sa->last_seq = seq; - } - else - { - pos = sa->last_seq - seq; - sa->replay_window |= (1ULL << pos); - } -} - -always_inline void -esp_replay_advance_esn (ipsec_sa_t * sa, u32 seq) -{ - int wrap = sa->seq_hi - sa->last_seq_hi; - u32 pos; - - if (wrap == 0 && seq > sa->last_seq) - { - pos = seq - sa->last_seq; - if (pos < ESP_WINDOW_SIZE) - sa->replay_window = ((sa->replay_window) << pos) | 1; - else - sa->replay_window = 1; - sa->last_seq = seq; - } - else if (wrap > 0) - { - pos = ~seq + sa->last_seq + 1; - if (pos < ESP_WINDOW_SIZE) - sa->replay_window = ((sa->replay_window) << pos) | 1; - else - sa->replay_window = 1; - sa->last_seq = seq; - sa->last_seq_hi = sa->seq_hi; - } - else if (wrap < 0) - { - pos = ~seq + sa->last_seq + 1; - sa->replay_window |= (1ULL << pos); - } - else - { - pos = sa->last_seq - seq; - sa->replay_window |= (1ULL << pos); - } -} - always_inline int esp_seq_advance (ipsec_sa_t * sa) { - if (PREDICT_TRUE (sa->use_esn)) + if (PREDICT_TRUE (ipsec_sa_is_set_USE_ESN (sa))) { if (PREDICT_FALSE (sa->seq == ESP_SEQ_MAX)) { - if (PREDICT_FALSE - (sa->use_anti_replay && sa->seq_hi == ESP_SEQ_MAX)) + if (PREDICT_FALSE (ipsec_sa_is_set_USE_ANTI_REPLAY (sa) && + sa->seq_hi == ESP_SEQ_MAX)) return 1; sa->seq_hi++; } @@ -195,7 +96,8 @@ esp_seq_advance (ipsec_sa_t * sa) } else { - if (PREDICT_FALSE (sa->use_anti_replay && sa->seq == ESP_SEQ_MAX)) + if (PREDICT_FALSE (ipsec_sa_is_set_USE_ANTI_REPLAY (sa) && + sa->seq == ESP_SEQ_MAX)) return 1; sa->seq++; } @@ -203,39 +105,29 @@ esp_seq_advance (ipsec_sa_t * sa) return 0; } - -always_inline unsigned int -hmac_calc (vlib_main_t * vm, ipsec_integ_alg_t alg, u8 * key, int key_len, - u8 * data, int data_len, u8 * signature, u8 use_esn, u32 seq_hi) +always_inline void +esp_aad_fill (vnet_crypto_op_t * op, + const esp_header_t * esp, const ipsec_sa_t * sa) { - ipsec_main_t *im = &ipsec_main; - vnet_crypto_op_t _op, *op = &_op; - ASSERT (alg < IPSEC_INTEG_N_ALG); - - if (PREDICT_FALSE (im->integ_algs[alg].op_type == 0)) - return 0; - - op->op = im->integ_algs[alg].op_type; - op->key = key; - op->key_len = key_len; - op->src = data; - op->len = data_len; - op->dst = signature; -#if 0 + esp_aead_t *aad; - HMAC_Init_ex (ctx, key, key_len, md, NULL); + aad = (esp_aead_t *) op->aad; + aad->data[0] = esp->spi; - HMAC_Update (ctx, data, data_len); - - if (PREDICT_TRUE (use_esn)) - HMAC_Update (ctx, (u8 *) & seq_hi, sizeof (seq_hi)); - HMAC_Final (ctx, signature, &len); - -#endif - vnet_crypto_process_ops (vm, op, 1); - return im->integ_algs[alg].trunc_size; + if (ipsec_sa_is_set_USE_ESN (sa)) + { + /* SPI, seq-hi, seq-low */ + aad->data[1] = clib_host_to_net_u32 (sa->seq_hi); + aad->data[2] = esp->seq; + op->aad_len = 12; + } + else + { + /* SPI, seq-low */ + aad->data[1] = esp->seq; + op->aad_len = 8; + } } - #endif /* __ESP_H__ */ /*