unittest_crypto_test_registration_t **r0 = a0;
unittest_crypto_test_registration_t **r1 = a1;
- return (r0[0]->op > r1[0]->op);
+ return (strncmp (r0[0]->name, r1[0]->name, 256));
}
static clib_error_t *
test_crypto (vlib_main_t * vm, crypto_test_main_t * tm)
{
+ vnet_crypto_main_t *cm = &crypto_main;
unittest_crypto_test_registration_t *r = tm->test_registrations;
unittest_crypto_test_registration_t **rv = 0;
+ vnet_crypto_alg_data_t *ad;
vnet_crypto_op_t *ops = 0, *op;
- u8 *computed_data = 0, *s = 0;
- u32 computed_data_total_len = 0, n_tests = 0;
+ vnet_crypto_key_index_t *key_indices = 0;
+ u8 *computed_data = 0, *s = 0, *err = 0;
+ u32 computed_data_total_len = 0, n_ops = 0;
u32 i;
/* construct registration vector */
while (r)
{
vec_add1 (rv, r);
- computed_data_total_len += r->data.length;
- n_tests += 1;
+ ad = vec_elt_at_index (cm->algs, r->alg);
+
+ for (i = 0; i < VNET_CRYPTO_OP_N_TYPES; i++)
+ {
+ vnet_crypto_op_id_t id = ad->op_by_type[i];
+
+ if (id == 0)
+ continue;
+
+ switch (i)
+ {
+ case VNET_CRYPTO_OP_TYPE_ENCRYPT:
+ case VNET_CRYPTO_OP_TYPE_DECRYPT:
+ case VNET_CRYPTO_OP_TYPE_AEAD_DECRYPT:
+ computed_data_total_len += r->ciphertext.length;
+ n_ops += 1;
+ break;
+ case VNET_CRYPTO_OP_TYPE_AEAD_ENCRYPT:
+ computed_data_total_len += r->ciphertext.length;
+ computed_data_total_len += r->tag.length;
+ n_ops += 1;
+ break;
+ case VNET_CRYPTO_OP_TYPE_HMAC:
+ computed_data_total_len += r->digest.length;
+ n_ops += 1;
+ break;
+ default:
+ break;
+ };
+ }
+
/* next */
r = r->next;
}
+ /* no tests registered */
+ if (n_ops == 0)
+ return 0;
+
vec_sort_with_function (rv, sort_registrations);
vec_validate_aligned (computed_data, computed_data_total_len - 1,
CLIB_CACHE_LINE_BYTES);
- vec_validate_aligned (ops, n_tests - 1, CLIB_CACHE_LINE_BYTES);
+ vec_validate_aligned (ops, n_ops - 1, CLIB_CACHE_LINE_BYTES);
computed_data_total_len = 0;
+ op = ops;
/* *INDENT-OFF* */
vec_foreach_index (i, rv)
{
r = rv[i];
- op = ops + i;
- op->op = r->op;
- op->iv = r->iv.data;
- op->key = r->key.data;
- op->src = r->data.data;
- op->dst = computed_data + computed_data_total_len;
- op->len = r->data.length;
- op->key_len = r->key.length;
- computed_data_total_len += r->expected.length;
- /* next */
- r = r->next;
+ int t;
+ ad = vec_elt_at_index (cm->algs, r->alg);
+ for (t = 0; t < VNET_CRYPTO_OP_N_TYPES; t++)
+ {
+ vnet_crypto_op_id_t id = ad->op_by_type[t];
+
+ if (id == 0)
+ continue;
+
+ vnet_crypto_op_init (op, id);
+
+ switch (t)
+ {
+ case VNET_CRYPTO_OP_TYPE_ENCRYPT:
+ case VNET_CRYPTO_OP_TYPE_DECRYPT:
+ op->iv = r->iv.data;
+ op->key_index = vnet_crypto_key_add (vm, r->alg,
+ r->key.data,
+ r->key.length);
+ vec_add1 (key_indices, op->key_index);
+ op->len = r->plaintext.length;
+ op->src = t == VNET_CRYPTO_OP_TYPE_ENCRYPT ?
+ r->plaintext.data : r->ciphertext.data;
+ op->dst = computed_data + computed_data_total_len;
+ computed_data_total_len += r->ciphertext.length;
+ break;
+ case VNET_CRYPTO_OP_TYPE_AEAD_ENCRYPT:
+ case VNET_CRYPTO_OP_TYPE_AEAD_DECRYPT:
+ op->iv = r->iv.data;
+ op->key_index = vnet_crypto_key_add (vm, r->alg,
+ r->key.data,
+ r->key.length);
+ vec_add1 (key_indices, op->key_index);
+ op->aad = r->aad.data;
+ op->aad_len = r->aad.length;
+ op->len = r->plaintext.length;
+ op->dst = computed_data + computed_data_total_len;
+ computed_data_total_len += r->ciphertext.length;
+ if (t == VNET_CRYPTO_OP_TYPE_AEAD_ENCRYPT)
+ {
+ op->src = r->plaintext.data;
+ op->tag = computed_data + computed_data_total_len;
+ computed_data_total_len += r->tag.length;
+ }
+ else
+ {
+ op->src = r->ciphertext.data;
+ op->tag = r->tag.data;
+ }
+ op->tag_len = r->tag.length;
+ break;
+ case VNET_CRYPTO_OP_TYPE_HMAC:
+ op->key_index = vnet_crypto_key_add (vm, r->alg,
+ r->key.data,
+ r->key.length);
+ vec_add1 (key_indices, op->key_index);
+ op->src = r->plaintext.data;
+ op->len = r->plaintext.length;
+ op->digest_len = r->digest.length;
+ op->digest = computed_data + computed_data_total_len;
+ computed_data_total_len += r->digest.length;
+ break;
+ default:
+ break;
+ };
+
+ op->user_data = i;
+ op++;
+ }
}
/* *INDENT-ON* */
vnet_crypto_process_ops (vm, ops, vec_len (ops));
/* *INDENT-OFF* */
- vec_foreach_index (i, rv)
+ vec_foreach (op, ops)
{
int fail = 0;
- r = rv[i];
- op = ops + i;
+ r = rv[op->user_data];
+ unittest_crypto_test_data_t *exp_pt = 0, *exp_ct = 0;
+ unittest_crypto_test_data_t *exp_digest = 0, *exp_tag = 0;
- if (memcmp (op->dst, r->expected.data, r->expected.length) != 0)
- fail = 1;
+ switch (vnet_crypto_get_op_type (op->op))
+ {
+ case VNET_CRYPTO_OP_TYPE_AEAD_ENCRYPT:
+ exp_tag = &r->tag;
+ /* fall through */
+ case VNET_CRYPTO_OP_TYPE_ENCRYPT:
+ exp_ct = &r->ciphertext;
+ break;
+ case VNET_CRYPTO_OP_TYPE_AEAD_DECRYPT:
+ case VNET_CRYPTO_OP_TYPE_DECRYPT:
+ exp_pt = &r->plaintext;
+ break;
+ case VNET_CRYPTO_OP_TYPE_HMAC:
+ exp_digest = &r->digest;
+ break;
+ default:
+ break;
+ }
+
+ vec_reset_length (err);
+
+ if (op->status != VNET_CRYPTO_OP_STATUS_COMPLETED)
+ err = format (err, "%sengine error: %U", vec_len (err) ? ", " : "",
+ format_vnet_crypto_op_status, op->status);
+
+ if (exp_ct && memcmp (op->dst, exp_ct->data, exp_ct->length) != 0)
+ err = format (err, "%sciphertext mismatch",
+ vec_len (err) ? ", " : "");
+
+ if (exp_pt && memcmp (op->dst, exp_pt->data, exp_pt->length) != 0)
+ err = format (err, "%splaintext mismatch", vec_len (err) ? ", " : "");
+
+ if (exp_tag && memcmp (op->tag, exp_tag->data, exp_tag->length) != 0)
+ err = format (err, "%stag mismatch", vec_len (err) ? ", " : "");
+
+ if (exp_digest &&
+ memcmp (op->digest, exp_digest->data, exp_digest->length) != 0)
+ err = format (err, "%sdigest mismatch", vec_len (err) ? ", " : "");
vec_reset_length (s);
- s = format (s, "%s (%U)", r->name,
- format_vnet_crypto_op, r->op);
+ s = format (s, "%s (%U)", r->name, format_vnet_crypto_op, op->op);
- vlib_cli_output (vm, "%-60v%s", s, fail ? "FAIL" : "OK");
- if (fail & tm->verbose)
+ if (vec_len (err))
+ fail = 1;
+
+ vlib_cli_output (vm, "%-60v%s%v", s, vec_len (err) ? "FAIL: " : "OK",
+ err);
+ if (tm->verbose)
{
- vlib_cli_output (vm, "Expected:\n%U\nCalculated:\n%U",
- format_hexdump, r->expected, r->expected.length,
- format_hexdump, op->dst, r->expected.length);
+ if (tm->verbose == 2)
+ fail = 1;
+
+ if (exp_ct && fail)
+ vlib_cli_output (vm, "Expected ciphertext:\n%U"
+ "\nCalculated ciphertext:\n%U",
+ format_hexdump, exp_ct->data, exp_ct->length,
+ format_hexdump, op->dst, exp_ct->length);
+ if (exp_pt && fail)
+ vlib_cli_output (vm, "Expected plaintext:\n%U"
+ "\nCalculated plaintext:\n%U",
+ format_hexdump, exp_pt->data, exp_pt->length,
+ format_hexdump, op->dst, exp_pt->length);
+ if (r->tag.length && fail)
+ vlib_cli_output (vm, "Expected tag:\n%U"
+ "\nCalculated tag:\n%U",
+ format_hexdump, r->tag.data, r->tag.length,
+ format_hexdump, op->tag, op->tag_len);
+ if (exp_digest && fail)
+ vlib_cli_output (vm, "Expected digest:\n%U"
+ "\nCalculated Digest:\n%U",
+ format_hexdump, exp_digest->data,
+ exp_digest->length, format_hexdump, op->digest,
+ op->digest_len);
}
}
+
+ vec_foreach_index (i, key_indices)
+ vnet_crypto_key_del (vm, key_indices[i]);
/* *INDENT-ON* */
vec_free (computed_data);
vec_free (ops);
+ vec_free (err);
vec_free (rv);
vec_free (s);
return 0;
}
+static u32
+test_crypto_get_key_sz (vnet_crypto_alg_t alg)
+{
+ switch (alg)
+ {
+#define _(n, s, l) \
+ case VNET_CRYPTO_ALG_##n: \
+ return l;
+ /* *INDENT-OFF* */
+ foreach_crypto_cipher_alg
+ foreach_crypto_aead_alg
+ /* *INDENT-ON* */
+#undef _
+ case VNET_CRYPTO_ALG_HMAC_MD5:
+ case VNET_CRYPTO_ALG_HMAC_SHA1:
+ return 20;
+ case VNET_CRYPTO_ALG_HMAC_SHA224:
+ return 28;
+ case VNET_CRYPTO_ALG_HMAC_SHA256:
+ return 32;
+ case VNET_CRYPTO_ALG_HMAC_SHA384:
+ return 48;
+ case VNET_CRYPTO_ALG_HMAC_SHA512:
+ return 64;
+ default:
+ return 0;
+ }
+
+ return 0;
+}
+
+static clib_error_t *
+test_crypto_perf (vlib_main_t * vm, crypto_test_main_t * tm)
+{
+ vnet_crypto_main_t *cm = &crypto_main;
+ clib_error_t *err = 0;
+ u32 n_buffers, n_alloc = 0, warmup_rounds, rounds;
+ u32 *buffer_indices = 0;
+ vnet_crypto_op_t *ops1 = 0, *ops2 = 0, *op1, *op2;
+ vnet_crypto_alg_data_t *ad = vec_elt_at_index (cm->algs, tm->alg);
+ vnet_crypto_key_index_t key_index = ~0;
+ u8 key[32];
+ int buffer_size = vlib_buffer_get_default_data_size (vm);
+ u64 seed = clib_cpu_time_now ();
+ u64 t0[5], t1[5], t2[5], n_bytes = 0;
+ int i, j;
+
+ if (tm->buffer_size > buffer_size)
+ return clib_error_return (0, "buffer size must be <= %u", buffer_size);
+
+ rounds = tm->rounds ? tm->rounds : 100;
+ n_buffers = tm->n_buffers ? tm->n_buffers : 256;
+ buffer_size = tm->buffer_size ? tm->buffer_size : 2048;
+ warmup_rounds = tm->warmup_rounds ? tm->warmup_rounds : 100;
+
+ if (buffer_size > vlib_buffer_get_default_data_size (vm))
+ return clib_error_return (0, "buffer size too big");
+
+ vec_validate_aligned (buffer_indices, n_buffers - 1, CLIB_CACHE_LINE_BYTES);
+ vec_validate_aligned (ops1, n_buffers - 1, CLIB_CACHE_LINE_BYTES);
+ vec_validate_aligned (ops2, n_buffers - 1, CLIB_CACHE_LINE_BYTES);
+
+ n_alloc = vlib_buffer_alloc (vm, buffer_indices, n_buffers);
+ if (n_alloc != n_buffers)
+ {
+ if (n_alloc)
+ vlib_buffer_free (vm, buffer_indices, n_alloc);
+ err = clib_error_return (0, "buffer alloc failure");
+ goto done;
+ }
+
+ vlib_cli_output (vm, "%U: n_buffers %u buffer-size %u rounds %u "
+ "warmup-rounds %u",
+ format_vnet_crypto_alg, tm->alg, n_buffers, buffer_size,
+ rounds, warmup_rounds);
+ vlib_cli_output (vm, " cpu-freq %.2f GHz",
+ (f64) vm->clib_time.clocks_per_second * 1e-9);
+
+ vnet_crypto_op_type_t ot = 0;
+
+ for (i = 0; i < sizeof (key); i++)
+ key[i] = i;
+
+ key_index = vnet_crypto_key_add (vm, tm->alg, key,
+ test_crypto_get_key_sz (tm->alg));
+
+ for (i = 0; i < VNET_CRYPTO_OP_N_TYPES; i++)
+ {
+ vnet_crypto_op_id_t id = ad->op_by_type[i];
+ if (id == 0)
+ continue;
+ ot = i;
+ break;
+ }
+
+ for (i = 0; i < n_buffers; i++)
+ {
+ vlib_buffer_t *b = vlib_get_buffer (vm, buffer_indices[i]);
+ op1 = ops1 + i;
+ op2 = ops2 + i;
+
+ switch (ot)
+ {
+ case VNET_CRYPTO_OP_TYPE_ENCRYPT:
+ case VNET_CRYPTO_OP_TYPE_DECRYPT:
+ vnet_crypto_op_init (op1,
+ ad->op_by_type[VNET_CRYPTO_OP_TYPE_ENCRYPT]);
+ vnet_crypto_op_init (op2,
+ ad->op_by_type[VNET_CRYPTO_OP_TYPE_DECRYPT]);
+ op1->flags = VNET_CRYPTO_OP_FLAG_INIT_IV;
+ op1->src = op2->src = op1->dst = op2->dst = b->data;
+ op1->key_index = op2->key_index = key_index;
+ op1->iv = op2->iv = b->data - 64;
+ n_bytes += op1->len = op2->len = buffer_size;
+ break;
+ case VNET_CRYPTO_OP_TYPE_AEAD_ENCRYPT:
+ case VNET_CRYPTO_OP_TYPE_AEAD_DECRYPT:
+ vnet_crypto_op_init (op1,
+ ad->op_by_type
+ [VNET_CRYPTO_OP_TYPE_AEAD_ENCRYPT]);
+ vnet_crypto_op_init (op2,
+ ad->op_by_type
+ [VNET_CRYPTO_OP_TYPE_AEAD_DECRYPT]);
+ op1->src = op2->src = op1->dst = op2->dst = b->data;
+ op1->key_index = op2->key_index = key_index;
+ op1->tag = op2->tag = b->data - 32;
+ op1->iv = op2->iv = b->data - 64;
+ op1->aad = op2->aad = b->data - VLIB_BUFFER_PRE_DATA_SIZE;
+ op1->aad_len = op2->aad_len = 64;
+ op1->tag_len = op2->tag_len = 16;
+ n_bytes += op1->len = op2->len = buffer_size;
+ break;
+ case VNET_CRYPTO_OP_TYPE_HMAC:
+ vnet_crypto_op_init (op1, ad->op_by_type[VNET_CRYPTO_OP_TYPE_HMAC]);
+ op1->src = b->data;
+ op1->key_index = key_index;
+ op1->iv = 0;
+ op1->digest = b->data - VLIB_BUFFER_PRE_DATA_SIZE;
+ op1->digest_len = 0;
+ n_bytes += op1->len = buffer_size;
+ break;
+ default:
+ return 0;
+ }
+
+ for (j = -VLIB_BUFFER_PRE_DATA_SIZE; j < buffer_size; j += 8)
+ *(u64 *) (b->data + j) = 1 + random_u64 (&seed);
+ }
+
+ for (i = 0; i < 5; i++)
+ {
+ for (j = 0; j < warmup_rounds; j++)
+ {
+ vnet_crypto_process_ops (vm, ops1, n_buffers);
+ if (ot != VNET_CRYPTO_OP_TYPE_HMAC)
+ vnet_crypto_process_ops (vm, ops2, n_buffers);
+ }
+
+ t0[i] = clib_cpu_time_now ();
+ for (j = 0; j < rounds; j++)
+ vnet_crypto_process_ops (vm, ops1, n_buffers);
+ t1[i] = clib_cpu_time_now ();
+
+ if (ot != VNET_CRYPTO_OP_TYPE_HMAC)
+ {
+ for (j = 0; j < rounds; j++)
+ vnet_crypto_process_ops (vm, ops2, n_buffers);
+ t2[i] = clib_cpu_time_now ();
+ }
+ }
+
+ for (i = 0; i < 5; i++)
+ {
+ f64 tpb1 = (f64) (t1[i] - t0[i]) / (n_bytes * rounds);
+ f64 gbps1 = vm->clib_time.clocks_per_second * 1e-9 * 8 / tpb1;
+ f64 tpb2, gbps2;
+
+ if (ot != VNET_CRYPTO_OP_TYPE_HMAC)
+ {
+ tpb2 = (f64) (t2[i] - t1[i]) / (n_bytes * rounds);
+ gbps2 = vm->clib_time.clocks_per_second * 1e-9 * 8 / tpb2;
+ vlib_cli_output (vm, "%-2u: encrypt %.03f ticks/byte, %.02f Gbps; "
+ "decrypt %.03f ticks/byte, %.02f Gbps",
+ i + 1, tpb1, gbps1, tpb2, gbps2);
+ }
+ else
+ {
+ vlib_cli_output (vm, "%-2u: hash %.03f ticks/byte, %.02f Gbps\n",
+ i + 1, tpb1, gbps1);
+ }
+ }
+
+done:
+ if (n_alloc)
+ vlib_buffer_free (vm, buffer_indices, n_alloc);
+
+ if (key_index != ~0)
+ vnet_crypto_key_del (vm, key_index);
+
+ vec_free (buffer_indices);
+ vec_free (ops1);
+ vec_free (ops2);
+ return err;
+}
+
static clib_error_t *
test_crypto_command_fn (vlib_main_t * vm,
unformat_input_t * input, vlib_cli_command_t * cmd)
{
crypto_test_main_t *tm = &crypto_test_main;
+ unittest_crypto_test_registration_t *tr;
+ int is_perf = 0;
+
+ tr = tm->test_registrations;
+ memset (tm, 0, sizeof (crypto_test_main_t));
+ tm->test_registrations = tr;
+ tm->alg = ~0;
while (unformat_check_input (input) != UNFORMAT_END_OF_INPUT)
{
- if (unformat (input, "verbose %d", &tm->verbose))
+ if (unformat (input, "verbose"))
+ tm->verbose = 1;
+ else if (unformat (input, "detail"))
+ tm->verbose = 2;
+ else
+ if (unformat (input, "perf %U", unformat_vnet_crypto_alg, &tm->alg))
+ is_perf = 1;
+ else if (unformat (input, "buffers %u", &tm->n_buffers))
+ ;
+ else if (unformat (input, "rounds %u", &tm->rounds))
+ ;
+ else if (unformat (input, "warmup-rounds %u", &tm->warmup_rounds))
+ ;
+ else if (unformat (input, "buffer-size %u", &tm->buffer_size))
;
else
return clib_error_return (0, "unknown input '%U'",
format_unformat_error, input);
}
- return test_crypto (vm, tm);
+ if (is_perf)
+ return test_crypto_perf (vm, tm);
+ else
+ return test_crypto (vm, tm);
}
/* *INDENT-OFF* */