f467c4ad23
Add support for the Adiantum encryption mode. Adiantum was designed by Paul Crowley and is specified by our paper: Adiantum: length-preserving encryption for entry-level processors (https://eprint.iacr.org/2018/720.pdf) See our paper for full details; this patch only provides an overview. Adiantum is a tweakable, length-preserving encryption mode designed for fast and secure disk encryption, especially on CPUs without dedicated crypto instructions. Adiantum encrypts each sector using the XChaCha12 stream cipher, two passes of an ε-almost-∆-universal (εA∆U) hash function, and an invocation of the AES-256 block cipher on a single 16-byte block. On CPUs without AES instructions, Adiantum is much faster than AES-XTS; for example, on ARM Cortex-A7, on 4096-byte sectors Adiantum encryption is about 4 times faster than AES-256-XTS encryption, and decryption about 5 times faster. Adiantum is a specialization of the more general HBSH construction. Our earlier proposal, HPolyC, was also a HBSH specialization, but it used a different εA∆U hash function, one based on Poly1305 only. Adiantum's εA∆U hash function, which is based primarily on the "NH" hash function like that used in UMAC (RFC4418), is about twice as fast as HPolyC's; consequently, Adiantum is about 20% faster than HPolyC. This speed comes with no loss of security: Adiantum is provably just as secure as HPolyC, in fact slightly *more* secure. Like HPolyC, Adiantum's security is reducible to that of XChaCha12 and AES-256, subject to a security bound. XChaCha12 itself has a security reduction to ChaCha12. Therefore, one need not "trust" Adiantum; one need only trust ChaCha12 and AES-256. Note that the εA∆U hash function is only used for its proven combinatorical properties so cannot be "broken". Adiantum is also a true wide-block encryption mode, so flipping any plaintext bit in the sector scrambles the entire ciphertext, and vice versa. No other such mode is available in the kernel currently; doing the same with XTS scrambles only 16 bytes. Adiantum also supports arbitrary-length tweaks and naturally supports any length input >= 16 bytes without needing "ciphertext stealing". For the stream cipher, Adiantum uses XChaCha12 rather than XChaCha20 in order to make encryption feasible on the widest range of devices. Although the 20-round variant is quite popular, the best known attacks on ChaCha are on only 7 rounds, so ChaCha12 still has a substantial security margin; in fact, larger than AES-256's. 12-round Salsa20 is also the eSTREAM recommendation. For the block cipher, Adiantum uses AES-256, despite it having a lower security margin than XChaCha12 and needing table lookups, due to AES's extensive adoption and analysis making it the obvious first choice. Nevertheless, for flexibility this patch also permits the "adiantum" template to be instantiated with XChaCha20 and/or with an alternate block cipher. We need Adiantum support in the kernel for use in dm-crypt and fscrypt, where currently the only other suitable options are block cipher modes such as AES-XTS. A big problem with this is that many low-end mobile devices (e.g. Android Go phones sold primarily in developing countries, as well as some smartwatches) still have CPUs that lack AES instructions, e.g. ARM Cortex-A7. Sadly, AES-XTS encryption is much too slow to be viable on these devices. We did find that some "lightweight" block ciphers are fast enough, but these suffer from problems such as not having much cryptanalysis or being too controversial. The ChaCha stream cipher has excellent performance but is insecure to use directly for disk encryption, since each sector's IV is reused each time it is overwritten. Even restricting the threat model to offline attacks only isn't enough, since modern flash storage devices don't guarantee that "overwrites" are really overwrites, due to wear-leveling. Adiantum avoids this problem by constructing a "tweakable super-pseudorandom permutation"; this is the strongest possible security model for length-preserving encryption. Of course, storing random nonces along with the ciphertext would be the ideal solution. But doing that with existing hardware and filesystems runs into major practical problems; in most cases it would require data journaling (like dm-integrity) which severely degrades performance. Thus, for now length-preserving encryption is still needed. Signed-off-by: Eric Biggers <ebiggers@google.com> Reviewed-by: Ard Biesheuvel <ard.biesheuvel@linaro.org> Signed-off-by: Herbert Xu <herbert@gondor.apana.org.au> (cherry picked from commit 059c2a4d8e164dccc3078e49e7f286023b019a98 https://git.kernel.org/pub/scm/linux/kernel/git/herbert/cryptodev-2.6.git master) Conflicts: crypto/tcrypt.c Bug: 112008522 Test: Among other things, I ran the relevant crypto self-tests: 1.) Build kernel with CONFIG_CRYPTO_MANAGER_DISABLE_TESTS *unset*, and all relevant crypto algorithms built-in, including: CONFIG_CRYPTO_ADIANTUM=y CONFIG_CRYPTO_CHACHA20=y CONFIG_CRYPTO_CHACHA20_NEON=y CONFIG_CRYPTO_NHPOLY1305=y CONFIG_CRYPTO_NHPOLY1305_NEON=y CONFIG_CRYPTO_POLY1305=y CONFIG_CRYPTO_AES=y CONFIG_CRYPTO_AES_ARM=y 2.) Boot and check dmesg for test failures. 3.) Instantiate "adiantum(xchacha12,aes)" and "adiantum(xchacha20,aes)" to trigger them to be tested. There are many ways to do this, but one way is to create a dm-crypt target that uses them, e.g. key=$(hexdump -n 32 -e '16/4 "%08X" 1 "\n"' /dev/urandom) dmsetup create crypt --table "0 $((1<<17)) crypt xchacha12,aes-adiantum-plain64 $key 0 /dev/vdc 0" dmsetup remove crypt dmsetup create crypt --table "0 $((1<<17)) crypt xchacha20,aes-adiantum-plain64 $key 0 /dev/vdc 0" dmsetup remove crypt 4.) Check dmesg for test failures again. 5.) Do 1-4 on both x86_64 (for basic testing) and on arm32 (for testing the ARM32-specific implementations). I did the arm32 kernel testing on Raspberry Pi 2, which is a BCM2836-based device that can run the upstream and Android common kernels. The same ARM32 assembly files for ChaCha, NHPoly1305, and AES are also included in the userspace Adiantum benchmark suite at https://github.com/google/adiantum, where they have undergone additional correctness testing. Change-Id: Ic61c13b53facfd2173065be715a7ee5f3af8760b Signed-off-by: Eric Biggers <ebiggers@google.com>
3015 lines
77 KiB
C
3015 lines
77 KiB
C
/*
|
|
* Quick & dirty crypto testing module.
|
|
*
|
|
* This will only exist until we have a better testing mechanism
|
|
* (e.g. a char device).
|
|
*
|
|
* Copyright (c) 2002 James Morris <jmorris@intercode.com.au>
|
|
* Copyright (c) 2002 Jean-Francois Dive <jef@linuxbe.org>
|
|
* Copyright (c) 2007 Nokia Siemens Networks
|
|
*
|
|
* Updated RFC4106 AES-GCM testing.
|
|
* Authors: Aidan O'Mahony (aidan.o.mahony@intel.com)
|
|
* Adrian Hoban <adrian.hoban@intel.com>
|
|
* Gabriele Paoloni <gabriele.paoloni@intel.com>
|
|
* Tadeusz Struk (tadeusz.struk@intel.com)
|
|
* Copyright (c) 2010, Intel Corporation.
|
|
*
|
|
* This program is free software; you can redistribute it and/or modify it
|
|
* under the terms of the GNU General Public License as published by the Free
|
|
* Software Foundation; either version 2 of the License, or (at your option)
|
|
* any later version.
|
|
*
|
|
*/
|
|
|
|
#define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
|
|
|
|
#include <crypto/aead.h>
|
|
#include <crypto/hash.h>
|
|
#include <crypto/skcipher.h>
|
|
#include <linux/err.h>
|
|
#include <linux/fips.h>
|
|
#include <linux/init.h>
|
|
#include <linux/gfp.h>
|
|
#include <linux/module.h>
|
|
#include <linux/scatterlist.h>
|
|
#include <linux/string.h>
|
|
#include <linux/moduleparam.h>
|
|
#include <linux/jiffies.h>
|
|
#include <linux/timex.h>
|
|
#include <linux/interrupt.h>
|
|
#include "tcrypt.h"
|
|
|
|
/*
|
|
* Need slab memory for testing (size in number of pages).
|
|
*/
|
|
#define TVMEMSIZE 4
|
|
|
|
/*
|
|
* Used by test_cipher_speed()
|
|
*/
|
|
#define ENCRYPT 1
|
|
#define DECRYPT 0
|
|
|
|
#define MAX_DIGEST_SIZE 64
|
|
|
|
/*
|
|
* return a string with the driver name
|
|
*/
|
|
#define get_driver_name(tfm_type, tfm) crypto_tfm_alg_driver_name(tfm_type ## _tfm(tfm))
|
|
|
|
/*
|
|
* Used by test_cipher_speed()
|
|
*/
|
|
static unsigned int sec;
|
|
|
|
static char *alg = NULL;
|
|
static u32 type;
|
|
static u32 mask;
|
|
static int mode;
|
|
static u32 num_mb = 8;
|
|
static char *tvmem[TVMEMSIZE];
|
|
|
|
static char *check[] = {
|
|
"des", "md5", "des3_ede", "rot13", "sha1", "sha224", "sha256", "sm3",
|
|
"blowfish", "twofish", "serpent", "sha384", "sha512", "md4", "aes",
|
|
"cast6", "arc4", "michael_mic", "deflate", "crc32c", "tea", "xtea",
|
|
"khazad", "wp512", "wp384", "wp256", "tnepres", "xeta", "fcrypt",
|
|
"camellia", "seed", "salsa20", "rmd128", "rmd160", "rmd256", "rmd320",
|
|
"lzo", "cts", "zlib", "sha3-224", "sha3-256", "sha3-384", "sha3-512",
|
|
NULL
|
|
};
|
|
|
|
static u32 block_sizes[] = { 16, 64, 256, 1024, 8192, 0 };
|
|
static u32 aead_sizes[] = { 16, 64, 256, 512, 1024, 2048, 4096, 8192, 0 };
|
|
|
|
#define XBUFSIZE 8
|
|
#define MAX_IVLEN 32
|
|
|
|
static int testmgr_alloc_buf(char *buf[XBUFSIZE])
|
|
{
|
|
int i;
|
|
|
|
for (i = 0; i < XBUFSIZE; i++) {
|
|
buf[i] = (void *)__get_free_page(GFP_KERNEL);
|
|
if (!buf[i])
|
|
goto err_free_buf;
|
|
}
|
|
|
|
return 0;
|
|
|
|
err_free_buf:
|
|
while (i-- > 0)
|
|
free_page((unsigned long)buf[i]);
|
|
|
|
return -ENOMEM;
|
|
}
|
|
|
|
static void testmgr_free_buf(char *buf[XBUFSIZE])
|
|
{
|
|
int i;
|
|
|
|
for (i = 0; i < XBUFSIZE; i++)
|
|
free_page((unsigned long)buf[i]);
|
|
}
|
|
|
|
static void sg_init_aead(struct scatterlist *sg, char *xbuf[XBUFSIZE],
|
|
unsigned int buflen, const void *assoc,
|
|
unsigned int aad_size)
|
|
{
|
|
int np = (buflen + PAGE_SIZE - 1)/PAGE_SIZE;
|
|
int k, rem;
|
|
|
|
if (np > XBUFSIZE) {
|
|
rem = PAGE_SIZE;
|
|
np = XBUFSIZE;
|
|
} else {
|
|
rem = buflen % PAGE_SIZE;
|
|
}
|
|
|
|
sg_init_table(sg, np + 1);
|
|
|
|
sg_set_buf(&sg[0], assoc, aad_size);
|
|
|
|
if (rem)
|
|
np--;
|
|
for (k = 0; k < np; k++)
|
|
sg_set_buf(&sg[k + 1], xbuf[k], PAGE_SIZE);
|
|
|
|
if (rem)
|
|
sg_set_buf(&sg[k + 1], xbuf[k], rem);
|
|
}
|
|
|
|
static inline int do_one_aead_op(struct aead_request *req, int ret)
|
|
{
|
|
struct crypto_wait *wait = req->base.data;
|
|
|
|
return crypto_wait_req(ret, wait);
|
|
}
|
|
|
|
struct test_mb_aead_data {
|
|
struct scatterlist sg[XBUFSIZE];
|
|
struct scatterlist sgout[XBUFSIZE];
|
|
struct aead_request *req;
|
|
struct crypto_wait wait;
|
|
char *xbuf[XBUFSIZE];
|
|
char *xoutbuf[XBUFSIZE];
|
|
char *axbuf[XBUFSIZE];
|
|
};
|
|
|
|
static int do_mult_aead_op(struct test_mb_aead_data *data, int enc,
|
|
u32 num_mb, int *rc)
|
|
{
|
|
int i, err = 0;
|
|
|
|
/* Fire up a bunch of concurrent requests */
|
|
for (i = 0; i < num_mb; i++) {
|
|
if (enc == ENCRYPT)
|
|
rc[i] = crypto_aead_encrypt(data[i].req);
|
|
else
|
|
rc[i] = crypto_aead_decrypt(data[i].req);
|
|
}
|
|
|
|
/* Wait for all requests to finish */
|
|
for (i = 0; i < num_mb; i++) {
|
|
rc[i] = crypto_wait_req(rc[i], &data[i].wait);
|
|
|
|
if (rc[i]) {
|
|
pr_info("concurrent request %d error %d\n", i, rc[i]);
|
|
err = rc[i];
|
|
}
|
|
}
|
|
|
|
return err;
|
|
}
|
|
|
|
static int test_mb_aead_jiffies(struct test_mb_aead_data *data, int enc,
|
|
int blen, int secs, u32 num_mb)
|
|
{
|
|
unsigned long start, end;
|
|
int bcount;
|
|
int ret = 0;
|
|
int *rc;
|
|
|
|
rc = kcalloc(num_mb, sizeof(*rc), GFP_KERNEL);
|
|
if (!rc)
|
|
return -ENOMEM;
|
|
|
|
for (start = jiffies, end = start + secs * HZ, bcount = 0;
|
|
time_before(jiffies, end); bcount++) {
|
|
ret = do_mult_aead_op(data, enc, num_mb, rc);
|
|
if (ret)
|
|
goto out;
|
|
}
|
|
|
|
pr_cont("%d operations in %d seconds (%ld bytes)\n",
|
|
bcount * num_mb, secs, (long)bcount * blen * num_mb);
|
|
|
|
out:
|
|
kfree(rc);
|
|
return ret;
|
|
}
|
|
|
|
static int test_mb_aead_cycles(struct test_mb_aead_data *data, int enc,
|
|
int blen, u32 num_mb)
|
|
{
|
|
unsigned long cycles = 0;
|
|
int ret = 0;
|
|
int i;
|
|
int *rc;
|
|
|
|
rc = kcalloc(num_mb, sizeof(*rc), GFP_KERNEL);
|
|
if (!rc)
|
|
return -ENOMEM;
|
|
|
|
/* Warm-up run. */
|
|
for (i = 0; i < 4; i++) {
|
|
ret = do_mult_aead_op(data, enc, num_mb, rc);
|
|
if (ret)
|
|
goto out;
|
|
}
|
|
|
|
/* The real thing. */
|
|
for (i = 0; i < 8; i++) {
|
|
cycles_t start, end;
|
|
|
|
start = get_cycles();
|
|
ret = do_mult_aead_op(data, enc, num_mb, rc);
|
|
end = get_cycles();
|
|
|
|
if (ret)
|
|
goto out;
|
|
|
|
cycles += end - start;
|
|
}
|
|
|
|
pr_cont("1 operation in %lu cycles (%d bytes)\n",
|
|
(cycles + 4) / (8 * num_mb), blen);
|
|
|
|
out:
|
|
kfree(rc);
|
|
return ret;
|
|
}
|
|
|
|
static void test_mb_aead_speed(const char *algo, int enc, int secs,
|
|
struct aead_speed_template *template,
|
|
unsigned int tcount, u8 authsize,
|
|
unsigned int aad_size, u8 *keysize, u32 num_mb)
|
|
{
|
|
struct test_mb_aead_data *data;
|
|
struct crypto_aead *tfm;
|
|
unsigned int i, j, iv_len;
|
|
const char *key;
|
|
const char *e;
|
|
void *assoc;
|
|
u32 *b_size;
|
|
char *iv;
|
|
int ret;
|
|
|
|
|
|
if (aad_size >= PAGE_SIZE) {
|
|
pr_err("associate data length (%u) too big\n", aad_size);
|
|
return;
|
|
}
|
|
|
|
iv = kzalloc(MAX_IVLEN, GFP_KERNEL);
|
|
if (!iv)
|
|
return;
|
|
|
|
if (enc == ENCRYPT)
|
|
e = "encryption";
|
|
else
|
|
e = "decryption";
|
|
|
|
data = kcalloc(num_mb, sizeof(*data), GFP_KERNEL);
|
|
if (!data)
|
|
goto out_free_iv;
|
|
|
|
tfm = crypto_alloc_aead(algo, 0, 0);
|
|
if (IS_ERR(tfm)) {
|
|
pr_err("failed to load transform for %s: %ld\n",
|
|
algo, PTR_ERR(tfm));
|
|
goto out_free_data;
|
|
}
|
|
|
|
ret = crypto_aead_setauthsize(tfm, authsize);
|
|
|
|
for (i = 0; i < num_mb; ++i)
|
|
if (testmgr_alloc_buf(data[i].xbuf)) {
|
|
while (i--)
|
|
testmgr_free_buf(data[i].xbuf);
|
|
goto out_free_tfm;
|
|
}
|
|
|
|
for (i = 0; i < num_mb; ++i)
|
|
if (testmgr_alloc_buf(data[i].axbuf)) {
|
|
while (i--)
|
|
testmgr_free_buf(data[i].axbuf);
|
|
goto out_free_xbuf;
|
|
}
|
|
|
|
for (i = 0; i < num_mb; ++i)
|
|
if (testmgr_alloc_buf(data[i].xoutbuf)) {
|
|
while (i--)
|
|
testmgr_free_buf(data[i].xoutbuf);
|
|
goto out_free_axbuf;
|
|
}
|
|
|
|
for (i = 0; i < num_mb; ++i) {
|
|
data[i].req = aead_request_alloc(tfm, GFP_KERNEL);
|
|
if (!data[i].req) {
|
|
pr_err("alg: skcipher: Failed to allocate request for %s\n",
|
|
algo);
|
|
while (i--)
|
|
aead_request_free(data[i].req);
|
|
goto out_free_xoutbuf;
|
|
}
|
|
}
|
|
|
|
for (i = 0; i < num_mb; ++i) {
|
|
crypto_init_wait(&data[i].wait);
|
|
aead_request_set_callback(data[i].req,
|
|
CRYPTO_TFM_REQ_MAY_BACKLOG,
|
|
crypto_req_done, &data[i].wait);
|
|
}
|
|
|
|
pr_info("\ntesting speed of multibuffer %s (%s) %s\n", algo,
|
|
get_driver_name(crypto_aead, tfm), e);
|
|
|
|
i = 0;
|
|
do {
|
|
b_size = aead_sizes;
|
|
do {
|
|
if (*b_size + authsize > XBUFSIZE * PAGE_SIZE) {
|
|
pr_err("template (%u) too big for buffer (%lu)\n",
|
|
authsize + *b_size,
|
|
XBUFSIZE * PAGE_SIZE);
|
|
goto out;
|
|
}
|
|
|
|
pr_info("test %u (%d bit key, %d byte blocks): ", i,
|
|
*keysize * 8, *b_size);
|
|
|
|
/* Set up tfm global state, i.e. the key */
|
|
|
|
memset(tvmem[0], 0xff, PAGE_SIZE);
|
|
key = tvmem[0];
|
|
for (j = 0; j < tcount; j++) {
|
|
if (template[j].klen == *keysize) {
|
|
key = template[j].key;
|
|
break;
|
|
}
|
|
}
|
|
|
|
crypto_aead_clear_flags(tfm, ~0);
|
|
|
|
ret = crypto_aead_setkey(tfm, key, *keysize);
|
|
if (ret) {
|
|
pr_err("setkey() failed flags=%x\n",
|
|
crypto_aead_get_flags(tfm));
|
|
goto out;
|
|
}
|
|
|
|
iv_len = crypto_aead_ivsize(tfm);
|
|
if (iv_len)
|
|
memset(iv, 0xff, iv_len);
|
|
|
|
/* Now setup per request stuff, i.e. buffers */
|
|
|
|
for (j = 0; j < num_mb; ++j) {
|
|
struct test_mb_aead_data *cur = &data[j];
|
|
|
|
assoc = cur->axbuf[0];
|
|
memset(assoc, 0xff, aad_size);
|
|
|
|
sg_init_aead(cur->sg, cur->xbuf,
|
|
*b_size + (enc ? 0 : authsize),
|
|
assoc, aad_size);
|
|
|
|
sg_init_aead(cur->sgout, cur->xoutbuf,
|
|
*b_size + (enc ? authsize : 0),
|
|
assoc, aad_size);
|
|
|
|
aead_request_set_ad(cur->req, aad_size);
|
|
|
|
if (!enc) {
|
|
|
|
aead_request_set_crypt(cur->req,
|
|
cur->sgout,
|
|
cur->sg,
|
|
*b_size, iv);
|
|
ret = crypto_aead_encrypt(cur->req);
|
|
ret = do_one_aead_op(cur->req, ret);
|
|
|
|
if (ret) {
|
|
pr_err("calculating auth failed failed (%d)\n",
|
|
ret);
|
|
break;
|
|
}
|
|
}
|
|
|
|
aead_request_set_crypt(cur->req, cur->sg,
|
|
cur->sgout, *b_size +
|
|
(enc ? 0 : authsize),
|
|
iv);
|
|
|
|
}
|
|
|
|
if (secs) {
|
|
ret = test_mb_aead_jiffies(data, enc, *b_size,
|
|
secs, num_mb);
|
|
cond_resched();
|
|
} else {
|
|
ret = test_mb_aead_cycles(data, enc, *b_size,
|
|
num_mb);
|
|
}
|
|
|
|
if (ret) {
|
|
pr_err("%s() failed return code=%d\n", e, ret);
|
|
break;
|
|
}
|
|
b_size++;
|
|
i++;
|
|
} while (*b_size);
|
|
keysize++;
|
|
} while (*keysize);
|
|
|
|
out:
|
|
for (i = 0; i < num_mb; ++i)
|
|
aead_request_free(data[i].req);
|
|
out_free_xoutbuf:
|
|
for (i = 0; i < num_mb; ++i)
|
|
testmgr_free_buf(data[i].xoutbuf);
|
|
out_free_axbuf:
|
|
for (i = 0; i < num_mb; ++i)
|
|
testmgr_free_buf(data[i].axbuf);
|
|
out_free_xbuf:
|
|
for (i = 0; i < num_mb; ++i)
|
|
testmgr_free_buf(data[i].xbuf);
|
|
out_free_tfm:
|
|
crypto_free_aead(tfm);
|
|
out_free_data:
|
|
kfree(data);
|
|
out_free_iv:
|
|
kfree(iv);
|
|
}
|
|
|
|
static int test_aead_jiffies(struct aead_request *req, int enc,
|
|
int blen, int secs)
|
|
{
|
|
unsigned long start, end;
|
|
int bcount;
|
|
int ret;
|
|
|
|
for (start = jiffies, end = start + secs * HZ, bcount = 0;
|
|
time_before(jiffies, end); bcount++) {
|
|
if (enc)
|
|
ret = do_one_aead_op(req, crypto_aead_encrypt(req));
|
|
else
|
|
ret = do_one_aead_op(req, crypto_aead_decrypt(req));
|
|
|
|
if (ret)
|
|
return ret;
|
|
}
|
|
|
|
printk("%d operations in %d seconds (%ld bytes)\n",
|
|
bcount, secs, (long)bcount * blen);
|
|
return 0;
|
|
}
|
|
|
|
static int test_aead_cycles(struct aead_request *req, int enc, int blen)
|
|
{
|
|
unsigned long cycles = 0;
|
|
int ret = 0;
|
|
int i;
|
|
|
|
/* Warm-up run. */
|
|
for (i = 0; i < 4; i++) {
|
|
if (enc)
|
|
ret = do_one_aead_op(req, crypto_aead_encrypt(req));
|
|
else
|
|
ret = do_one_aead_op(req, crypto_aead_decrypt(req));
|
|
|
|
if (ret)
|
|
goto out;
|
|
}
|
|
|
|
/* The real thing. */
|
|
for (i = 0; i < 8; i++) {
|
|
cycles_t start, end;
|
|
|
|
start = get_cycles();
|
|
if (enc)
|
|
ret = do_one_aead_op(req, crypto_aead_encrypt(req));
|
|
else
|
|
ret = do_one_aead_op(req, crypto_aead_decrypt(req));
|
|
end = get_cycles();
|
|
|
|
if (ret)
|
|
goto out;
|
|
|
|
cycles += end - start;
|
|
}
|
|
|
|
out:
|
|
if (ret == 0)
|
|
printk("1 operation in %lu cycles (%d bytes)\n",
|
|
(cycles + 4) / 8, blen);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static void test_aead_speed(const char *algo, int enc, unsigned int secs,
|
|
struct aead_speed_template *template,
|
|
unsigned int tcount, u8 authsize,
|
|
unsigned int aad_size, u8 *keysize)
|
|
{
|
|
unsigned int i, j;
|
|
struct crypto_aead *tfm;
|
|
int ret = -ENOMEM;
|
|
const char *key;
|
|
struct aead_request *req;
|
|
struct scatterlist *sg;
|
|
struct scatterlist *sgout;
|
|
const char *e;
|
|
void *assoc;
|
|
char *iv;
|
|
char *xbuf[XBUFSIZE];
|
|
char *xoutbuf[XBUFSIZE];
|
|
char *axbuf[XBUFSIZE];
|
|
unsigned int *b_size;
|
|
unsigned int iv_len;
|
|
struct crypto_wait wait;
|
|
|
|
iv = kzalloc(MAX_IVLEN, GFP_KERNEL);
|
|
if (!iv)
|
|
return;
|
|
|
|
if (aad_size >= PAGE_SIZE) {
|
|
pr_err("associate data length (%u) too big\n", aad_size);
|
|
goto out_noxbuf;
|
|
}
|
|
|
|
if (enc == ENCRYPT)
|
|
e = "encryption";
|
|
else
|
|
e = "decryption";
|
|
|
|
if (testmgr_alloc_buf(xbuf))
|
|
goto out_noxbuf;
|
|
if (testmgr_alloc_buf(axbuf))
|
|
goto out_noaxbuf;
|
|
if (testmgr_alloc_buf(xoutbuf))
|
|
goto out_nooutbuf;
|
|
|
|
sg = kmalloc(sizeof(*sg) * 9 * 2, GFP_KERNEL);
|
|
if (!sg)
|
|
goto out_nosg;
|
|
sgout = &sg[9];
|
|
|
|
tfm = crypto_alloc_aead(algo, 0, 0);
|
|
|
|
if (IS_ERR(tfm)) {
|
|
pr_err("alg: aead: Failed to load transform for %s: %ld\n", algo,
|
|
PTR_ERR(tfm));
|
|
goto out_notfm;
|
|
}
|
|
|
|
crypto_init_wait(&wait);
|
|
printk(KERN_INFO "\ntesting speed of %s (%s) %s\n", algo,
|
|
get_driver_name(crypto_aead, tfm), e);
|
|
|
|
req = aead_request_alloc(tfm, GFP_KERNEL);
|
|
if (!req) {
|
|
pr_err("alg: aead: Failed to allocate request for %s\n",
|
|
algo);
|
|
goto out_noreq;
|
|
}
|
|
|
|
aead_request_set_callback(req, CRYPTO_TFM_REQ_MAY_BACKLOG,
|
|
crypto_req_done, &wait);
|
|
|
|
i = 0;
|
|
do {
|
|
b_size = aead_sizes;
|
|
do {
|
|
assoc = axbuf[0];
|
|
memset(assoc, 0xff, aad_size);
|
|
|
|
if ((*keysize + *b_size) > TVMEMSIZE * PAGE_SIZE) {
|
|
pr_err("template (%u) too big for tvmem (%lu)\n",
|
|
*keysize + *b_size,
|
|
TVMEMSIZE * PAGE_SIZE);
|
|
goto out;
|
|
}
|
|
|
|
key = tvmem[0];
|
|
for (j = 0; j < tcount; j++) {
|
|
if (template[j].klen == *keysize) {
|
|
key = template[j].key;
|
|
break;
|
|
}
|
|
}
|
|
ret = crypto_aead_setkey(tfm, key, *keysize);
|
|
ret = crypto_aead_setauthsize(tfm, authsize);
|
|
|
|
iv_len = crypto_aead_ivsize(tfm);
|
|
if (iv_len)
|
|
memset(iv, 0xff, iv_len);
|
|
|
|
crypto_aead_clear_flags(tfm, ~0);
|
|
printk(KERN_INFO "test %u (%d bit key, %d byte blocks): ",
|
|
i, *keysize * 8, *b_size);
|
|
|
|
|
|
memset(tvmem[0], 0xff, PAGE_SIZE);
|
|
|
|
if (ret) {
|
|
pr_err("setkey() failed flags=%x\n",
|
|
crypto_aead_get_flags(tfm));
|
|
goto out;
|
|
}
|
|
|
|
sg_init_aead(sg, xbuf, *b_size + (enc ? 0 : authsize),
|
|
assoc, aad_size);
|
|
|
|
sg_init_aead(sgout, xoutbuf,
|
|
*b_size + (enc ? authsize : 0), assoc,
|
|
aad_size);
|
|
|
|
aead_request_set_ad(req, aad_size);
|
|
|
|
if (!enc) {
|
|
|
|
/*
|
|
* For decryption we need a proper auth so
|
|
* we do the encryption path once with buffers
|
|
* reversed (input <-> output) to calculate it
|
|
*/
|
|
aead_request_set_crypt(req, sgout, sg,
|
|
*b_size, iv);
|
|
ret = do_one_aead_op(req,
|
|
crypto_aead_encrypt(req));
|
|
|
|
if (ret) {
|
|
pr_err("calculating auth failed failed (%d)\n",
|
|
ret);
|
|
break;
|
|
}
|
|
}
|
|
|
|
aead_request_set_crypt(req, sg, sgout,
|
|
*b_size + (enc ? 0 : authsize),
|
|
iv);
|
|
|
|
if (secs) {
|
|
ret = test_aead_jiffies(req, enc, *b_size,
|
|
secs);
|
|
cond_resched();
|
|
} else {
|
|
ret = test_aead_cycles(req, enc, *b_size);
|
|
}
|
|
|
|
if (ret) {
|
|
pr_err("%s() failed return code=%d\n", e, ret);
|
|
break;
|
|
}
|
|
b_size++;
|
|
i++;
|
|
} while (*b_size);
|
|
keysize++;
|
|
} while (*keysize);
|
|
|
|
out:
|
|
aead_request_free(req);
|
|
out_noreq:
|
|
crypto_free_aead(tfm);
|
|
out_notfm:
|
|
kfree(sg);
|
|
out_nosg:
|
|
testmgr_free_buf(xoutbuf);
|
|
out_nooutbuf:
|
|
testmgr_free_buf(axbuf);
|
|
out_noaxbuf:
|
|
testmgr_free_buf(xbuf);
|
|
out_noxbuf:
|
|
kfree(iv);
|
|
}
|
|
|
|
static void test_hash_sg_init(struct scatterlist *sg)
|
|
{
|
|
int i;
|
|
|
|
sg_init_table(sg, TVMEMSIZE);
|
|
for (i = 0; i < TVMEMSIZE; i++) {
|
|
sg_set_buf(sg + i, tvmem[i], PAGE_SIZE);
|
|
memset(tvmem[i], 0xff, PAGE_SIZE);
|
|
}
|
|
}
|
|
|
|
static inline int do_one_ahash_op(struct ahash_request *req, int ret)
|
|
{
|
|
struct crypto_wait *wait = req->base.data;
|
|
|
|
return crypto_wait_req(ret, wait);
|
|
}
|
|
|
|
struct test_mb_ahash_data {
|
|
struct scatterlist sg[XBUFSIZE];
|
|
char result[64];
|
|
struct ahash_request *req;
|
|
struct crypto_wait wait;
|
|
char *xbuf[XBUFSIZE];
|
|
};
|
|
|
|
static inline int do_mult_ahash_op(struct test_mb_ahash_data *data, u32 num_mb,
|
|
int *rc)
|
|
{
|
|
int i, err = 0;
|
|
|
|
/* Fire up a bunch of concurrent requests */
|
|
for (i = 0; i < num_mb; i++)
|
|
rc[i] = crypto_ahash_digest(data[i].req);
|
|
|
|
/* Wait for all requests to finish */
|
|
for (i = 0; i < num_mb; i++) {
|
|
rc[i] = crypto_wait_req(rc[i], &data[i].wait);
|
|
|
|
if (rc[i]) {
|
|
pr_info("concurrent request %d error %d\n", i, rc[i]);
|
|
err = rc[i];
|
|
}
|
|
}
|
|
|
|
return err;
|
|
}
|
|
|
|
static int test_mb_ahash_jiffies(struct test_mb_ahash_data *data, int blen,
|
|
int secs, u32 num_mb)
|
|
{
|
|
unsigned long start, end;
|
|
int bcount;
|
|
int ret = 0;
|
|
int *rc;
|
|
|
|
rc = kcalloc(num_mb, sizeof(*rc), GFP_KERNEL);
|
|
if (!rc)
|
|
return -ENOMEM;
|
|
|
|
for (start = jiffies, end = start + secs * HZ, bcount = 0;
|
|
time_before(jiffies, end); bcount++) {
|
|
ret = do_mult_ahash_op(data, num_mb, rc);
|
|
if (ret)
|
|
goto out;
|
|
}
|
|
|
|
pr_cont("%d operations in %d seconds (%ld bytes)\n",
|
|
bcount * num_mb, secs, (long)bcount * blen * num_mb);
|
|
|
|
out:
|
|
kfree(rc);
|
|
return ret;
|
|
}
|
|
|
|
static int test_mb_ahash_cycles(struct test_mb_ahash_data *data, int blen,
|
|
u32 num_mb)
|
|
{
|
|
unsigned long cycles = 0;
|
|
int ret = 0;
|
|
int i;
|
|
int *rc;
|
|
|
|
rc = kcalloc(num_mb, sizeof(*rc), GFP_KERNEL);
|
|
if (!rc)
|
|
return -ENOMEM;
|
|
|
|
/* Warm-up run. */
|
|
for (i = 0; i < 4; i++) {
|
|
ret = do_mult_ahash_op(data, num_mb, rc);
|
|
if (ret)
|
|
goto out;
|
|
}
|
|
|
|
/* The real thing. */
|
|
for (i = 0; i < 8; i++) {
|
|
cycles_t start, end;
|
|
|
|
start = get_cycles();
|
|
ret = do_mult_ahash_op(data, num_mb, rc);
|
|
end = get_cycles();
|
|
|
|
if (ret)
|
|
goto out;
|
|
|
|
cycles += end - start;
|
|
}
|
|
|
|
pr_cont("1 operation in %lu cycles (%d bytes)\n",
|
|
(cycles + 4) / (8 * num_mb), blen);
|
|
|
|
out:
|
|
kfree(rc);
|
|
return ret;
|
|
}
|
|
|
|
static void test_mb_ahash_speed(const char *algo, unsigned int secs,
|
|
struct hash_speed *speed, u32 num_mb)
|
|
{
|
|
struct test_mb_ahash_data *data;
|
|
struct crypto_ahash *tfm;
|
|
unsigned int i, j, k;
|
|
int ret;
|
|
|
|
data = kcalloc(num_mb, sizeof(*data), GFP_KERNEL);
|
|
if (!data)
|
|
return;
|
|
|
|
tfm = crypto_alloc_ahash(algo, 0, 0);
|
|
if (IS_ERR(tfm)) {
|
|
pr_err("failed to load transform for %s: %ld\n",
|
|
algo, PTR_ERR(tfm));
|
|
goto free_data;
|
|
}
|
|
|
|
for (i = 0; i < num_mb; ++i) {
|
|
if (testmgr_alloc_buf(data[i].xbuf))
|
|
goto out;
|
|
|
|
crypto_init_wait(&data[i].wait);
|
|
|
|
data[i].req = ahash_request_alloc(tfm, GFP_KERNEL);
|
|
if (!data[i].req) {
|
|
pr_err("alg: hash: Failed to allocate request for %s\n",
|
|
algo);
|
|
goto out;
|
|
}
|
|
|
|
ahash_request_set_callback(data[i].req, 0, crypto_req_done,
|
|
&data[i].wait);
|
|
|
|
sg_init_table(data[i].sg, XBUFSIZE);
|
|
for (j = 0; j < XBUFSIZE; j++) {
|
|
sg_set_buf(data[i].sg + j, data[i].xbuf[j], PAGE_SIZE);
|
|
memset(data[i].xbuf[j], 0xff, PAGE_SIZE);
|
|
}
|
|
}
|
|
|
|
pr_info("\ntesting speed of multibuffer %s (%s)\n", algo,
|
|
get_driver_name(crypto_ahash, tfm));
|
|
|
|
for (i = 0; speed[i].blen != 0; i++) {
|
|
/* For some reason this only tests digests. */
|
|
if (speed[i].blen != speed[i].plen)
|
|
continue;
|
|
|
|
if (speed[i].blen > XBUFSIZE * PAGE_SIZE) {
|
|
pr_err("template (%u) too big for tvmem (%lu)\n",
|
|
speed[i].blen, XBUFSIZE * PAGE_SIZE);
|
|
goto out;
|
|
}
|
|
|
|
if (speed[i].klen)
|
|
crypto_ahash_setkey(tfm, tvmem[0], speed[i].klen);
|
|
|
|
for (k = 0; k < num_mb; k++)
|
|
ahash_request_set_crypt(data[k].req, data[k].sg,
|
|
data[k].result, speed[i].blen);
|
|
|
|
pr_info("test%3u "
|
|
"(%5u byte blocks,%5u bytes per update,%4u updates): ",
|
|
i, speed[i].blen, speed[i].plen,
|
|
speed[i].blen / speed[i].plen);
|
|
|
|
if (secs) {
|
|
ret = test_mb_ahash_jiffies(data, speed[i].blen, secs,
|
|
num_mb);
|
|
cond_resched();
|
|
} else {
|
|
ret = test_mb_ahash_cycles(data, speed[i].blen, num_mb);
|
|
}
|
|
|
|
|
|
if (ret) {
|
|
pr_err("At least one hashing failed ret=%d\n", ret);
|
|
break;
|
|
}
|
|
}
|
|
|
|
out:
|
|
for (k = 0; k < num_mb; ++k)
|
|
ahash_request_free(data[k].req);
|
|
|
|
for (k = 0; k < num_mb; ++k)
|
|
testmgr_free_buf(data[k].xbuf);
|
|
|
|
crypto_free_ahash(tfm);
|
|
|
|
free_data:
|
|
kfree(data);
|
|
}
|
|
|
|
static int test_ahash_jiffies_digest(struct ahash_request *req, int blen,
|
|
char *out, int secs)
|
|
{
|
|
unsigned long start, end;
|
|
int bcount;
|
|
int ret;
|
|
|
|
for (start = jiffies, end = start + secs * HZ, bcount = 0;
|
|
time_before(jiffies, end); bcount++) {
|
|
ret = do_one_ahash_op(req, crypto_ahash_digest(req));
|
|
if (ret)
|
|
return ret;
|
|
}
|
|
|
|
printk("%6u opers/sec, %9lu bytes/sec\n",
|
|
bcount / secs, ((long)bcount * blen) / secs);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int test_ahash_jiffies(struct ahash_request *req, int blen,
|
|
int plen, char *out, int secs)
|
|
{
|
|
unsigned long start, end;
|
|
int bcount, pcount;
|
|
int ret;
|
|
|
|
if (plen == blen)
|
|
return test_ahash_jiffies_digest(req, blen, out, secs);
|
|
|
|
for (start = jiffies, end = start + secs * HZ, bcount = 0;
|
|
time_before(jiffies, end); bcount++) {
|
|
ret = do_one_ahash_op(req, crypto_ahash_init(req));
|
|
if (ret)
|
|
return ret;
|
|
for (pcount = 0; pcount < blen; pcount += plen) {
|
|
ret = do_one_ahash_op(req, crypto_ahash_update(req));
|
|
if (ret)
|
|
return ret;
|
|
}
|
|
/* we assume there is enough space in 'out' for the result */
|
|
ret = do_one_ahash_op(req, crypto_ahash_final(req));
|
|
if (ret)
|
|
return ret;
|
|
}
|
|
|
|
pr_cont("%6u opers/sec, %9lu bytes/sec\n",
|
|
bcount / secs, ((long)bcount * blen) / secs);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int test_ahash_cycles_digest(struct ahash_request *req, int blen,
|
|
char *out)
|
|
{
|
|
unsigned long cycles = 0;
|
|
int ret, i;
|
|
|
|
/* Warm-up run. */
|
|
for (i = 0; i < 4; i++) {
|
|
ret = do_one_ahash_op(req, crypto_ahash_digest(req));
|
|
if (ret)
|
|
goto out;
|
|
}
|
|
|
|
/* The real thing. */
|
|
for (i = 0; i < 8; i++) {
|
|
cycles_t start, end;
|
|
|
|
start = get_cycles();
|
|
|
|
ret = do_one_ahash_op(req, crypto_ahash_digest(req));
|
|
if (ret)
|
|
goto out;
|
|
|
|
end = get_cycles();
|
|
|
|
cycles += end - start;
|
|
}
|
|
|
|
out:
|
|
if (ret)
|
|
return ret;
|
|
|
|
pr_cont("%6lu cycles/operation, %4lu cycles/byte\n",
|
|
cycles / 8, cycles / (8 * blen));
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int test_ahash_cycles(struct ahash_request *req, int blen,
|
|
int plen, char *out)
|
|
{
|
|
unsigned long cycles = 0;
|
|
int i, pcount, ret;
|
|
|
|
if (plen == blen)
|
|
return test_ahash_cycles_digest(req, blen, out);
|
|
|
|
/* Warm-up run. */
|
|
for (i = 0; i < 4; i++) {
|
|
ret = do_one_ahash_op(req, crypto_ahash_init(req));
|
|
if (ret)
|
|
goto out;
|
|
for (pcount = 0; pcount < blen; pcount += plen) {
|
|
ret = do_one_ahash_op(req, crypto_ahash_update(req));
|
|
if (ret)
|
|
goto out;
|
|
}
|
|
ret = do_one_ahash_op(req, crypto_ahash_final(req));
|
|
if (ret)
|
|
goto out;
|
|
}
|
|
|
|
/* The real thing. */
|
|
for (i = 0; i < 8; i++) {
|
|
cycles_t start, end;
|
|
|
|
start = get_cycles();
|
|
|
|
ret = do_one_ahash_op(req, crypto_ahash_init(req));
|
|
if (ret)
|
|
goto out;
|
|
for (pcount = 0; pcount < blen; pcount += plen) {
|
|
ret = do_one_ahash_op(req, crypto_ahash_update(req));
|
|
if (ret)
|
|
goto out;
|
|
}
|
|
ret = do_one_ahash_op(req, crypto_ahash_final(req));
|
|
if (ret)
|
|
goto out;
|
|
|
|
end = get_cycles();
|
|
|
|
cycles += end - start;
|
|
}
|
|
|
|
out:
|
|
if (ret)
|
|
return ret;
|
|
|
|
pr_cont("%6lu cycles/operation, %4lu cycles/byte\n",
|
|
cycles / 8, cycles / (8 * blen));
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void test_ahash_speed_common(const char *algo, unsigned int secs,
|
|
struct hash_speed *speed, unsigned mask)
|
|
{
|
|
struct scatterlist sg[TVMEMSIZE];
|
|
struct crypto_wait wait;
|
|
struct ahash_request *req;
|
|
struct crypto_ahash *tfm;
|
|
char *output;
|
|
int i, ret;
|
|
|
|
tfm = crypto_alloc_ahash(algo, 0, mask);
|
|
if (IS_ERR(tfm)) {
|
|
pr_err("failed to load transform for %s: %ld\n",
|
|
algo, PTR_ERR(tfm));
|
|
return;
|
|
}
|
|
|
|
printk(KERN_INFO "\ntesting speed of async %s (%s)\n", algo,
|
|
get_driver_name(crypto_ahash, tfm));
|
|
|
|
if (crypto_ahash_digestsize(tfm) > MAX_DIGEST_SIZE) {
|
|
pr_err("digestsize(%u) > %d\n", crypto_ahash_digestsize(tfm),
|
|
MAX_DIGEST_SIZE);
|
|
goto out;
|
|
}
|
|
|
|
test_hash_sg_init(sg);
|
|
req = ahash_request_alloc(tfm, GFP_KERNEL);
|
|
if (!req) {
|
|
pr_err("ahash request allocation failure\n");
|
|
goto out;
|
|
}
|
|
|
|
crypto_init_wait(&wait);
|
|
ahash_request_set_callback(req, CRYPTO_TFM_REQ_MAY_BACKLOG,
|
|
crypto_req_done, &wait);
|
|
|
|
output = kmalloc(MAX_DIGEST_SIZE, GFP_KERNEL);
|
|
if (!output)
|
|
goto out_nomem;
|
|
|
|
for (i = 0; speed[i].blen != 0; i++) {
|
|
if (speed[i].blen > TVMEMSIZE * PAGE_SIZE) {
|
|
pr_err("template (%u) too big for tvmem (%lu)\n",
|
|
speed[i].blen, TVMEMSIZE * PAGE_SIZE);
|
|
break;
|
|
}
|
|
|
|
if (speed[i].klen)
|
|
crypto_ahash_setkey(tfm, tvmem[0], speed[i].klen);
|
|
|
|
pr_info("test%3u "
|
|
"(%5u byte blocks,%5u bytes per update,%4u updates): ",
|
|
i, speed[i].blen, speed[i].plen, speed[i].blen / speed[i].plen);
|
|
|
|
ahash_request_set_crypt(req, sg, output, speed[i].plen);
|
|
|
|
if (secs) {
|
|
ret = test_ahash_jiffies(req, speed[i].blen,
|
|
speed[i].plen, output, secs);
|
|
cond_resched();
|
|
} else {
|
|
ret = test_ahash_cycles(req, speed[i].blen,
|
|
speed[i].plen, output);
|
|
}
|
|
|
|
if (ret) {
|
|
pr_err("hashing failed ret=%d\n", ret);
|
|
break;
|
|
}
|
|
}
|
|
|
|
kfree(output);
|
|
|
|
out_nomem:
|
|
ahash_request_free(req);
|
|
|
|
out:
|
|
crypto_free_ahash(tfm);
|
|
}
|
|
|
|
static void test_ahash_speed(const char *algo, unsigned int secs,
|
|
struct hash_speed *speed)
|
|
{
|
|
return test_ahash_speed_common(algo, secs, speed, 0);
|
|
}
|
|
|
|
static void test_hash_speed(const char *algo, unsigned int secs,
|
|
struct hash_speed *speed)
|
|
{
|
|
return test_ahash_speed_common(algo, secs, speed, CRYPTO_ALG_ASYNC);
|
|
}
|
|
|
|
struct test_mb_skcipher_data {
|
|
struct scatterlist sg[XBUFSIZE];
|
|
struct skcipher_request *req;
|
|
struct crypto_wait wait;
|
|
char *xbuf[XBUFSIZE];
|
|
};
|
|
|
|
static int do_mult_acipher_op(struct test_mb_skcipher_data *data, int enc,
|
|
u32 num_mb, int *rc)
|
|
{
|
|
int i, err = 0;
|
|
|
|
/* Fire up a bunch of concurrent requests */
|
|
for (i = 0; i < num_mb; i++) {
|
|
if (enc == ENCRYPT)
|
|
rc[i] = crypto_skcipher_encrypt(data[i].req);
|
|
else
|
|
rc[i] = crypto_skcipher_decrypt(data[i].req);
|
|
}
|
|
|
|
/* Wait for all requests to finish */
|
|
for (i = 0; i < num_mb; i++) {
|
|
rc[i] = crypto_wait_req(rc[i], &data[i].wait);
|
|
|
|
if (rc[i]) {
|
|
pr_info("concurrent request %d error %d\n", i, rc[i]);
|
|
err = rc[i];
|
|
}
|
|
}
|
|
|
|
return err;
|
|
}
|
|
|
|
static int test_mb_acipher_jiffies(struct test_mb_skcipher_data *data, int enc,
|
|
int blen, int secs, u32 num_mb)
|
|
{
|
|
unsigned long start, end;
|
|
int bcount;
|
|
int ret = 0;
|
|
int *rc;
|
|
|
|
rc = kcalloc(num_mb, sizeof(*rc), GFP_KERNEL);
|
|
if (!rc)
|
|
return -ENOMEM;
|
|
|
|
for (start = jiffies, end = start + secs * HZ, bcount = 0;
|
|
time_before(jiffies, end); bcount++) {
|
|
ret = do_mult_acipher_op(data, enc, num_mb, rc);
|
|
if (ret)
|
|
goto out;
|
|
}
|
|
|
|
pr_cont("%d operations in %d seconds (%ld bytes)\n",
|
|
bcount * num_mb, secs, (long)bcount * blen * num_mb);
|
|
|
|
out:
|
|
kfree(rc);
|
|
return ret;
|
|
}
|
|
|
|
static int test_mb_acipher_cycles(struct test_mb_skcipher_data *data, int enc,
|
|
int blen, u32 num_mb)
|
|
{
|
|
unsigned long cycles = 0;
|
|
int ret = 0;
|
|
int i;
|
|
int *rc;
|
|
|
|
rc = kcalloc(num_mb, sizeof(*rc), GFP_KERNEL);
|
|
if (!rc)
|
|
return -ENOMEM;
|
|
|
|
/* Warm-up run. */
|
|
for (i = 0; i < 4; i++) {
|
|
ret = do_mult_acipher_op(data, enc, num_mb, rc);
|
|
if (ret)
|
|
goto out;
|
|
}
|
|
|
|
/* The real thing. */
|
|
for (i = 0; i < 8; i++) {
|
|
cycles_t start, end;
|
|
|
|
start = get_cycles();
|
|
ret = do_mult_acipher_op(data, enc, num_mb, rc);
|
|
end = get_cycles();
|
|
|
|
if (ret)
|
|
goto out;
|
|
|
|
cycles += end - start;
|
|
}
|
|
|
|
pr_cont("1 operation in %lu cycles (%d bytes)\n",
|
|
(cycles + 4) / (8 * num_mb), blen);
|
|
|
|
out:
|
|
kfree(rc);
|
|
return ret;
|
|
}
|
|
|
|
static void test_mb_skcipher_speed(const char *algo, int enc, int secs,
|
|
struct cipher_speed_template *template,
|
|
unsigned int tcount, u8 *keysize, u32 num_mb)
|
|
{
|
|
struct test_mb_skcipher_data *data;
|
|
struct crypto_skcipher *tfm;
|
|
unsigned int i, j, iv_len;
|
|
const char *key;
|
|
const char *e;
|
|
u32 *b_size;
|
|
char iv[128];
|
|
int ret;
|
|
|
|
if (enc == ENCRYPT)
|
|
e = "encryption";
|
|
else
|
|
e = "decryption";
|
|
|
|
data = kcalloc(num_mb, sizeof(*data), GFP_KERNEL);
|
|
if (!data)
|
|
return;
|
|
|
|
tfm = crypto_alloc_skcipher(algo, 0, 0);
|
|
if (IS_ERR(tfm)) {
|
|
pr_err("failed to load transform for %s: %ld\n",
|
|
algo, PTR_ERR(tfm));
|
|
goto out_free_data;
|
|
}
|
|
|
|
for (i = 0; i < num_mb; ++i)
|
|
if (testmgr_alloc_buf(data[i].xbuf)) {
|
|
while (i--)
|
|
testmgr_free_buf(data[i].xbuf);
|
|
goto out_free_tfm;
|
|
}
|
|
|
|
|
|
for (i = 0; i < num_mb; ++i)
|
|
if (testmgr_alloc_buf(data[i].xbuf)) {
|
|
while (i--)
|
|
testmgr_free_buf(data[i].xbuf);
|
|
goto out_free_tfm;
|
|
}
|
|
|
|
|
|
for (i = 0; i < num_mb; ++i) {
|
|
data[i].req = skcipher_request_alloc(tfm, GFP_KERNEL);
|
|
if (!data[i].req) {
|
|
pr_err("alg: skcipher: Failed to allocate request for %s\n",
|
|
algo);
|
|
while (i--)
|
|
skcipher_request_free(data[i].req);
|
|
goto out_free_xbuf;
|
|
}
|
|
}
|
|
|
|
for (i = 0; i < num_mb; ++i) {
|
|
skcipher_request_set_callback(data[i].req,
|
|
CRYPTO_TFM_REQ_MAY_BACKLOG,
|
|
crypto_req_done, &data[i].wait);
|
|
crypto_init_wait(&data[i].wait);
|
|
}
|
|
|
|
pr_info("\ntesting speed of multibuffer %s (%s) %s\n", algo,
|
|
get_driver_name(crypto_skcipher, tfm), e);
|
|
|
|
i = 0;
|
|
do {
|
|
b_size = block_sizes;
|
|
do {
|
|
if (*b_size > XBUFSIZE * PAGE_SIZE) {
|
|
pr_err("template (%u) too big for buffer (%lu)\n",
|
|
*b_size, XBUFSIZE * PAGE_SIZE);
|
|
goto out;
|
|
}
|
|
|
|
pr_info("test %u (%d bit key, %d byte blocks): ", i,
|
|
*keysize * 8, *b_size);
|
|
|
|
/* Set up tfm global state, i.e. the key */
|
|
|
|
memset(tvmem[0], 0xff, PAGE_SIZE);
|
|
key = tvmem[0];
|
|
for (j = 0; j < tcount; j++) {
|
|
if (template[j].klen == *keysize) {
|
|
key = template[j].key;
|
|
break;
|
|
}
|
|
}
|
|
|
|
crypto_skcipher_clear_flags(tfm, ~0);
|
|
|
|
ret = crypto_skcipher_setkey(tfm, key, *keysize);
|
|
if (ret) {
|
|
pr_err("setkey() failed flags=%x\n",
|
|
crypto_skcipher_get_flags(tfm));
|
|
goto out;
|
|
}
|
|
|
|
iv_len = crypto_skcipher_ivsize(tfm);
|
|
if (iv_len)
|
|
memset(&iv, 0xff, iv_len);
|
|
|
|
/* Now setup per request stuff, i.e. buffers */
|
|
|
|
for (j = 0; j < num_mb; ++j) {
|
|
struct test_mb_skcipher_data *cur = &data[j];
|
|
unsigned int k = *b_size;
|
|
unsigned int pages = DIV_ROUND_UP(k, PAGE_SIZE);
|
|
unsigned int p = 0;
|
|
|
|
sg_init_table(cur->sg, pages);
|
|
|
|
while (k > PAGE_SIZE) {
|
|
sg_set_buf(cur->sg + p, cur->xbuf[p],
|
|
PAGE_SIZE);
|
|
memset(cur->xbuf[p], 0xff, PAGE_SIZE);
|
|
p++;
|
|
k -= PAGE_SIZE;
|
|
}
|
|
|
|
sg_set_buf(cur->sg + p, cur->xbuf[p], k);
|
|
memset(cur->xbuf[p], 0xff, k);
|
|
|
|
skcipher_request_set_crypt(cur->req, cur->sg,
|
|
cur->sg, *b_size,
|
|
iv);
|
|
}
|
|
|
|
if (secs) {
|
|
ret = test_mb_acipher_jiffies(data, enc,
|
|
*b_size, secs,
|
|
num_mb);
|
|
cond_resched();
|
|
} else {
|
|
ret = test_mb_acipher_cycles(data, enc,
|
|
*b_size, num_mb);
|
|
}
|
|
|
|
if (ret) {
|
|
pr_err("%s() failed flags=%x\n", e,
|
|
crypto_skcipher_get_flags(tfm));
|
|
break;
|
|
}
|
|
b_size++;
|
|
i++;
|
|
} while (*b_size);
|
|
keysize++;
|
|
} while (*keysize);
|
|
|
|
out:
|
|
for (i = 0; i < num_mb; ++i)
|
|
skcipher_request_free(data[i].req);
|
|
out_free_xbuf:
|
|
for (i = 0; i < num_mb; ++i)
|
|
testmgr_free_buf(data[i].xbuf);
|
|
out_free_tfm:
|
|
crypto_free_skcipher(tfm);
|
|
out_free_data:
|
|
kfree(data);
|
|
}
|
|
|
|
static inline int do_one_acipher_op(struct skcipher_request *req, int ret)
|
|
{
|
|
struct crypto_wait *wait = req->base.data;
|
|
|
|
return crypto_wait_req(ret, wait);
|
|
}
|
|
|
|
static int test_acipher_jiffies(struct skcipher_request *req, int enc,
|
|
int blen, int secs)
|
|
{
|
|
unsigned long start, end;
|
|
int bcount;
|
|
int ret;
|
|
|
|
for (start = jiffies, end = start + secs * HZ, bcount = 0;
|
|
time_before(jiffies, end); bcount++) {
|
|
if (enc)
|
|
ret = do_one_acipher_op(req,
|
|
crypto_skcipher_encrypt(req));
|
|
else
|
|
ret = do_one_acipher_op(req,
|
|
crypto_skcipher_decrypt(req));
|
|
|
|
if (ret)
|
|
return ret;
|
|
}
|
|
|
|
pr_cont("%d operations in %d seconds (%ld bytes)\n",
|
|
bcount, secs, (long)bcount * blen);
|
|
return 0;
|
|
}
|
|
|
|
static int test_acipher_cycles(struct skcipher_request *req, int enc,
|
|
int blen)
|
|
{
|
|
unsigned long cycles = 0;
|
|
int ret = 0;
|
|
int i;
|
|
|
|
/* Warm-up run. */
|
|
for (i = 0; i < 4; i++) {
|
|
if (enc)
|
|
ret = do_one_acipher_op(req,
|
|
crypto_skcipher_encrypt(req));
|
|
else
|
|
ret = do_one_acipher_op(req,
|
|
crypto_skcipher_decrypt(req));
|
|
|
|
if (ret)
|
|
goto out;
|
|
}
|
|
|
|
/* The real thing. */
|
|
for (i = 0; i < 8; i++) {
|
|
cycles_t start, end;
|
|
|
|
start = get_cycles();
|
|
if (enc)
|
|
ret = do_one_acipher_op(req,
|
|
crypto_skcipher_encrypt(req));
|
|
else
|
|
ret = do_one_acipher_op(req,
|
|
crypto_skcipher_decrypt(req));
|
|
end = get_cycles();
|
|
|
|
if (ret)
|
|
goto out;
|
|
|
|
cycles += end - start;
|
|
}
|
|
|
|
out:
|
|
if (ret == 0)
|
|
pr_cont("1 operation in %lu cycles (%d bytes)\n",
|
|
(cycles + 4) / 8, blen);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static void test_skcipher_speed(const char *algo, int enc, unsigned int secs,
|
|
struct cipher_speed_template *template,
|
|
unsigned int tcount, u8 *keysize, bool async)
|
|
{
|
|
unsigned int ret, i, j, k, iv_len;
|
|
struct crypto_wait wait;
|
|
const char *key;
|
|
char iv[128];
|
|
struct skcipher_request *req;
|
|
struct crypto_skcipher *tfm;
|
|
const char *e;
|
|
u32 *b_size;
|
|
|
|
if (enc == ENCRYPT)
|
|
e = "encryption";
|
|
else
|
|
e = "decryption";
|
|
|
|
crypto_init_wait(&wait);
|
|
|
|
tfm = crypto_alloc_skcipher(algo, 0, async ? 0 : CRYPTO_ALG_ASYNC);
|
|
|
|
if (IS_ERR(tfm)) {
|
|
pr_err("failed to load transform for %s: %ld\n", algo,
|
|
PTR_ERR(tfm));
|
|
return;
|
|
}
|
|
|
|
pr_info("\ntesting speed of async %s (%s) %s\n", algo,
|
|
get_driver_name(crypto_skcipher, tfm), e);
|
|
|
|
req = skcipher_request_alloc(tfm, GFP_KERNEL);
|
|
if (!req) {
|
|
pr_err("tcrypt: skcipher: Failed to allocate request for %s\n",
|
|
algo);
|
|
goto out;
|
|
}
|
|
|
|
skcipher_request_set_callback(req, CRYPTO_TFM_REQ_MAY_BACKLOG,
|
|
crypto_req_done, &wait);
|
|
|
|
i = 0;
|
|
do {
|
|
b_size = block_sizes;
|
|
|
|
do {
|
|
struct scatterlist sg[TVMEMSIZE];
|
|
|
|
if ((*keysize + *b_size) > TVMEMSIZE * PAGE_SIZE) {
|
|
pr_err("template (%u) too big for "
|
|
"tvmem (%lu)\n", *keysize + *b_size,
|
|
TVMEMSIZE * PAGE_SIZE);
|
|
goto out_free_req;
|
|
}
|
|
|
|
pr_info("test %u (%d bit key, %d byte blocks): ", i,
|
|
*keysize * 8, *b_size);
|
|
|
|
memset(tvmem[0], 0xff, PAGE_SIZE);
|
|
|
|
/* set key, plain text and IV */
|
|
key = tvmem[0];
|
|
for (j = 0; j < tcount; j++) {
|
|
if (template[j].klen == *keysize) {
|
|
key = template[j].key;
|
|
break;
|
|
}
|
|
}
|
|
|
|
crypto_skcipher_clear_flags(tfm, ~0);
|
|
|
|
ret = crypto_skcipher_setkey(tfm, key, *keysize);
|
|
if (ret) {
|
|
pr_err("setkey() failed flags=%x\n",
|
|
crypto_skcipher_get_flags(tfm));
|
|
goto out_free_req;
|
|
}
|
|
|
|
k = *keysize + *b_size;
|
|
sg_init_table(sg, DIV_ROUND_UP(k, PAGE_SIZE));
|
|
|
|
if (k > PAGE_SIZE) {
|
|
sg_set_buf(sg, tvmem[0] + *keysize,
|
|
PAGE_SIZE - *keysize);
|
|
k -= PAGE_SIZE;
|
|
j = 1;
|
|
while (k > PAGE_SIZE) {
|
|
sg_set_buf(sg + j, tvmem[j], PAGE_SIZE);
|
|
memset(tvmem[j], 0xff, PAGE_SIZE);
|
|
j++;
|
|
k -= PAGE_SIZE;
|
|
}
|
|
sg_set_buf(sg + j, tvmem[j], k);
|
|
memset(tvmem[j], 0xff, k);
|
|
} else {
|
|
sg_set_buf(sg, tvmem[0] + *keysize, *b_size);
|
|
}
|
|
|
|
iv_len = crypto_skcipher_ivsize(tfm);
|
|
if (iv_len)
|
|
memset(&iv, 0xff, iv_len);
|
|
|
|
skcipher_request_set_crypt(req, sg, sg, *b_size, iv);
|
|
|
|
if (secs) {
|
|
ret = test_acipher_jiffies(req, enc,
|
|
*b_size, secs);
|
|
cond_resched();
|
|
} else {
|
|
ret = test_acipher_cycles(req, enc,
|
|
*b_size);
|
|
}
|
|
|
|
if (ret) {
|
|
pr_err("%s() failed flags=%x\n", e,
|
|
crypto_skcipher_get_flags(tfm));
|
|
break;
|
|
}
|
|
b_size++;
|
|
i++;
|
|
} while (*b_size);
|
|
keysize++;
|
|
} while (*keysize);
|
|
|
|
out_free_req:
|
|
skcipher_request_free(req);
|
|
out:
|
|
crypto_free_skcipher(tfm);
|
|
}
|
|
|
|
static void test_acipher_speed(const char *algo, int enc, unsigned int secs,
|
|
struct cipher_speed_template *template,
|
|
unsigned int tcount, u8 *keysize)
|
|
{
|
|
return test_skcipher_speed(algo, enc, secs, template, tcount, keysize,
|
|
true);
|
|
}
|
|
|
|
static void test_cipher_speed(const char *algo, int enc, unsigned int secs,
|
|
struct cipher_speed_template *template,
|
|
unsigned int tcount, u8 *keysize)
|
|
{
|
|
return test_skcipher_speed(algo, enc, secs, template, tcount, keysize,
|
|
false);
|
|
}
|
|
|
|
static void test_available(void)
|
|
{
|
|
char **name = check;
|
|
|
|
while (*name) {
|
|
printk("alg %s ", *name);
|
|
printk(crypto_has_alg(*name, 0, 0) ?
|
|
"found\n" : "not found\n");
|
|
name++;
|
|
}
|
|
}
|
|
|
|
static inline int tcrypt_test(const char *alg)
|
|
{
|
|
int ret;
|
|
|
|
pr_debug("testing %s\n", alg);
|
|
|
|
ret = alg_test(alg, alg, 0, 0);
|
|
/* non-fips algs return -EINVAL in fips mode */
|
|
if (fips_enabled && ret == -EINVAL)
|
|
ret = 0;
|
|
return ret;
|
|
}
|
|
|
|
static int do_test(const char *alg, u32 type, u32 mask, int m, u32 num_mb)
|
|
{
|
|
int i;
|
|
int ret = 0;
|
|
|
|
switch (m) {
|
|
case 0:
|
|
if (alg) {
|
|
if (!crypto_has_alg(alg, type,
|
|
mask ?: CRYPTO_ALG_TYPE_MASK))
|
|
ret = -ENOENT;
|
|
break;
|
|
}
|
|
|
|
for (i = 1; i < 200; i++)
|
|
ret += do_test(NULL, 0, 0, i, num_mb);
|
|
break;
|
|
|
|
case 1:
|
|
ret += tcrypt_test("md5");
|
|
break;
|
|
|
|
case 2:
|
|
ret += tcrypt_test("sha1");
|
|
break;
|
|
|
|
case 3:
|
|
ret += tcrypt_test("ecb(des)");
|
|
ret += tcrypt_test("cbc(des)");
|
|
ret += tcrypt_test("ctr(des)");
|
|
break;
|
|
|
|
case 4:
|
|
ret += tcrypt_test("ecb(des3_ede)");
|
|
ret += tcrypt_test("cbc(des3_ede)");
|
|
ret += tcrypt_test("ctr(des3_ede)");
|
|
break;
|
|
|
|
case 5:
|
|
ret += tcrypt_test("md4");
|
|
break;
|
|
|
|
case 6:
|
|
ret += tcrypt_test("sha256");
|
|
break;
|
|
|
|
case 7:
|
|
ret += tcrypt_test("ecb(blowfish)");
|
|
ret += tcrypt_test("cbc(blowfish)");
|
|
ret += tcrypt_test("ctr(blowfish)");
|
|
break;
|
|
|
|
case 8:
|
|
ret += tcrypt_test("ecb(twofish)");
|
|
ret += tcrypt_test("cbc(twofish)");
|
|
ret += tcrypt_test("ctr(twofish)");
|
|
ret += tcrypt_test("lrw(twofish)");
|
|
ret += tcrypt_test("xts(twofish)");
|
|
break;
|
|
|
|
case 9:
|
|
ret += tcrypt_test("ecb(serpent)");
|
|
ret += tcrypt_test("cbc(serpent)");
|
|
ret += tcrypt_test("ctr(serpent)");
|
|
ret += tcrypt_test("lrw(serpent)");
|
|
ret += tcrypt_test("xts(serpent)");
|
|
break;
|
|
|
|
case 10:
|
|
ret += tcrypt_test("ecb(aes)");
|
|
ret += tcrypt_test("cbc(aes)");
|
|
ret += tcrypt_test("lrw(aes)");
|
|
ret += tcrypt_test("xts(aes)");
|
|
ret += tcrypt_test("ctr(aes)");
|
|
ret += tcrypt_test("rfc3686(ctr(aes))");
|
|
break;
|
|
|
|
case 11:
|
|
ret += tcrypt_test("sha384");
|
|
break;
|
|
|
|
case 12:
|
|
ret += tcrypt_test("sha512");
|
|
break;
|
|
|
|
case 13:
|
|
ret += tcrypt_test("deflate");
|
|
break;
|
|
|
|
case 14:
|
|
ret += tcrypt_test("ecb(cast5)");
|
|
ret += tcrypt_test("cbc(cast5)");
|
|
ret += tcrypt_test("ctr(cast5)");
|
|
break;
|
|
|
|
case 15:
|
|
ret += tcrypt_test("ecb(cast6)");
|
|
ret += tcrypt_test("cbc(cast6)");
|
|
ret += tcrypt_test("ctr(cast6)");
|
|
ret += tcrypt_test("lrw(cast6)");
|
|
ret += tcrypt_test("xts(cast6)");
|
|
break;
|
|
|
|
case 16:
|
|
ret += tcrypt_test("ecb(arc4)");
|
|
break;
|
|
|
|
case 17:
|
|
ret += tcrypt_test("michael_mic");
|
|
break;
|
|
|
|
case 18:
|
|
ret += tcrypt_test("crc32c");
|
|
break;
|
|
|
|
case 19:
|
|
ret += tcrypt_test("ecb(tea)");
|
|
break;
|
|
|
|
case 20:
|
|
ret += tcrypt_test("ecb(xtea)");
|
|
break;
|
|
|
|
case 21:
|
|
ret += tcrypt_test("ecb(khazad)");
|
|
break;
|
|
|
|
case 22:
|
|
ret += tcrypt_test("wp512");
|
|
break;
|
|
|
|
case 23:
|
|
ret += tcrypt_test("wp384");
|
|
break;
|
|
|
|
case 24:
|
|
ret += tcrypt_test("wp256");
|
|
break;
|
|
|
|
case 25:
|
|
ret += tcrypt_test("ecb(tnepres)");
|
|
break;
|
|
|
|
case 26:
|
|
ret += tcrypt_test("ecb(anubis)");
|
|
ret += tcrypt_test("cbc(anubis)");
|
|
break;
|
|
|
|
case 27:
|
|
ret += tcrypt_test("tgr192");
|
|
break;
|
|
|
|
case 28:
|
|
ret += tcrypt_test("tgr160");
|
|
break;
|
|
|
|
case 29:
|
|
ret += tcrypt_test("tgr128");
|
|
break;
|
|
|
|
case 30:
|
|
ret += tcrypt_test("ecb(xeta)");
|
|
break;
|
|
|
|
case 31:
|
|
ret += tcrypt_test("pcbc(fcrypt)");
|
|
break;
|
|
|
|
case 32:
|
|
ret += tcrypt_test("ecb(camellia)");
|
|
ret += tcrypt_test("cbc(camellia)");
|
|
ret += tcrypt_test("ctr(camellia)");
|
|
ret += tcrypt_test("lrw(camellia)");
|
|
ret += tcrypt_test("xts(camellia)");
|
|
break;
|
|
|
|
case 33:
|
|
ret += tcrypt_test("sha224");
|
|
break;
|
|
|
|
case 34:
|
|
ret += tcrypt_test("salsa20");
|
|
break;
|
|
|
|
case 35:
|
|
ret += tcrypt_test("gcm(aes)");
|
|
break;
|
|
|
|
case 36:
|
|
ret += tcrypt_test("lzo");
|
|
break;
|
|
|
|
case 37:
|
|
ret += tcrypt_test("ccm(aes)");
|
|
break;
|
|
|
|
case 38:
|
|
ret += tcrypt_test("cts(cbc(aes))");
|
|
break;
|
|
|
|
case 39:
|
|
ret += tcrypt_test("rmd128");
|
|
break;
|
|
|
|
case 40:
|
|
ret += tcrypt_test("rmd160");
|
|
break;
|
|
|
|
case 41:
|
|
ret += tcrypt_test("rmd256");
|
|
break;
|
|
|
|
case 42:
|
|
ret += tcrypt_test("rmd320");
|
|
break;
|
|
|
|
case 43:
|
|
ret += tcrypt_test("ecb(seed)");
|
|
break;
|
|
|
|
case 44:
|
|
ret += tcrypt_test("zlib");
|
|
break;
|
|
|
|
case 45:
|
|
ret += tcrypt_test("rfc4309(ccm(aes))");
|
|
break;
|
|
|
|
case 46:
|
|
ret += tcrypt_test("ghash");
|
|
break;
|
|
|
|
case 47:
|
|
ret += tcrypt_test("crct10dif");
|
|
break;
|
|
|
|
case 48:
|
|
ret += tcrypt_test("sha3-224");
|
|
break;
|
|
|
|
case 49:
|
|
ret += tcrypt_test("sha3-256");
|
|
break;
|
|
|
|
case 50:
|
|
ret += tcrypt_test("sha3-384");
|
|
break;
|
|
|
|
case 51:
|
|
ret += tcrypt_test("sha3-512");
|
|
break;
|
|
|
|
case 52:
|
|
ret += tcrypt_test("sm3");
|
|
break;
|
|
|
|
case 100:
|
|
ret += tcrypt_test("hmac(md5)");
|
|
break;
|
|
|
|
case 101:
|
|
ret += tcrypt_test("hmac(sha1)");
|
|
break;
|
|
|
|
case 102:
|
|
ret += tcrypt_test("hmac(sha256)");
|
|
break;
|
|
|
|
case 103:
|
|
ret += tcrypt_test("hmac(sha384)");
|
|
break;
|
|
|
|
case 104:
|
|
ret += tcrypt_test("hmac(sha512)");
|
|
break;
|
|
|
|
case 105:
|
|
ret += tcrypt_test("hmac(sha224)");
|
|
break;
|
|
|
|
case 106:
|
|
ret += tcrypt_test("xcbc(aes)");
|
|
break;
|
|
|
|
case 107:
|
|
ret += tcrypt_test("hmac(rmd128)");
|
|
break;
|
|
|
|
case 108:
|
|
ret += tcrypt_test("hmac(rmd160)");
|
|
break;
|
|
|
|
case 109:
|
|
ret += tcrypt_test("vmac64(aes)");
|
|
break;
|
|
|
|
case 111:
|
|
ret += tcrypt_test("hmac(sha3-224)");
|
|
break;
|
|
|
|
case 112:
|
|
ret += tcrypt_test("hmac(sha3-256)");
|
|
break;
|
|
|
|
case 113:
|
|
ret += tcrypt_test("hmac(sha3-384)");
|
|
break;
|
|
|
|
case 114:
|
|
ret += tcrypt_test("hmac(sha3-512)");
|
|
break;
|
|
|
|
case 150:
|
|
ret += tcrypt_test("ansi_cprng");
|
|
break;
|
|
|
|
case 151:
|
|
ret += tcrypt_test("rfc4106(gcm(aes))");
|
|
break;
|
|
|
|
case 152:
|
|
ret += tcrypt_test("rfc4543(gcm(aes))");
|
|
break;
|
|
|
|
case 153:
|
|
ret += tcrypt_test("cmac(aes)");
|
|
break;
|
|
|
|
case 154:
|
|
ret += tcrypt_test("cmac(des3_ede)");
|
|
break;
|
|
|
|
case 155:
|
|
ret += tcrypt_test("authenc(hmac(sha1),cbc(aes))");
|
|
break;
|
|
|
|
case 156:
|
|
ret += tcrypt_test("authenc(hmac(md5),ecb(cipher_null))");
|
|
break;
|
|
|
|
case 157:
|
|
ret += tcrypt_test("authenc(hmac(sha1),ecb(cipher_null))");
|
|
break;
|
|
case 181:
|
|
ret += tcrypt_test("authenc(hmac(sha1),cbc(des))");
|
|
break;
|
|
case 182:
|
|
ret += tcrypt_test("authenc(hmac(sha1),cbc(des3_ede))");
|
|
break;
|
|
case 183:
|
|
ret += tcrypt_test("authenc(hmac(sha224),cbc(des))");
|
|
break;
|
|
case 184:
|
|
ret += tcrypt_test("authenc(hmac(sha224),cbc(des3_ede))");
|
|
break;
|
|
case 185:
|
|
ret += tcrypt_test("authenc(hmac(sha256),cbc(des))");
|
|
break;
|
|
case 186:
|
|
ret += tcrypt_test("authenc(hmac(sha256),cbc(des3_ede))");
|
|
break;
|
|
case 187:
|
|
ret += tcrypt_test("authenc(hmac(sha384),cbc(des))");
|
|
break;
|
|
case 188:
|
|
ret += tcrypt_test("authenc(hmac(sha384),cbc(des3_ede))");
|
|
break;
|
|
case 189:
|
|
ret += tcrypt_test("authenc(hmac(sha512),cbc(des))");
|
|
break;
|
|
case 190:
|
|
ret += tcrypt_test("authenc(hmac(sha512),cbc(des3_ede))");
|
|
break;
|
|
case 191:
|
|
ret += tcrypt_test("ecb(sm4)");
|
|
break;
|
|
case 200:
|
|
test_cipher_speed("ecb(aes)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_16_24_32);
|
|
test_cipher_speed("ecb(aes)", DECRYPT, sec, NULL, 0,
|
|
speed_template_16_24_32);
|
|
test_cipher_speed("cbc(aes)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_16_24_32);
|
|
test_cipher_speed("cbc(aes)", DECRYPT, sec, NULL, 0,
|
|
speed_template_16_24_32);
|
|
test_cipher_speed("lrw(aes)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_32_40_48);
|
|
test_cipher_speed("lrw(aes)", DECRYPT, sec, NULL, 0,
|
|
speed_template_32_40_48);
|
|
test_cipher_speed("xts(aes)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_32_64);
|
|
test_cipher_speed("xts(aes)", DECRYPT, sec, NULL, 0,
|
|
speed_template_32_64);
|
|
test_cipher_speed("cts(cbc(aes))", ENCRYPT, sec, NULL, 0,
|
|
speed_template_16_24_32);
|
|
test_cipher_speed("cts(cbc(aes))", DECRYPT, sec, NULL, 0,
|
|
speed_template_16_24_32);
|
|
test_cipher_speed("ctr(aes)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_16_24_32);
|
|
test_cipher_speed("ctr(aes)", DECRYPT, sec, NULL, 0,
|
|
speed_template_16_24_32);
|
|
break;
|
|
|
|
case 201:
|
|
test_cipher_speed("ecb(des3_ede)", ENCRYPT, sec,
|
|
des3_speed_template, DES3_SPEED_VECTORS,
|
|
speed_template_24);
|
|
test_cipher_speed("ecb(des3_ede)", DECRYPT, sec,
|
|
des3_speed_template, DES3_SPEED_VECTORS,
|
|
speed_template_24);
|
|
test_cipher_speed("cbc(des3_ede)", ENCRYPT, sec,
|
|
des3_speed_template, DES3_SPEED_VECTORS,
|
|
speed_template_24);
|
|
test_cipher_speed("cbc(des3_ede)", DECRYPT, sec,
|
|
des3_speed_template, DES3_SPEED_VECTORS,
|
|
speed_template_24);
|
|
test_cipher_speed("ctr(des3_ede)", ENCRYPT, sec,
|
|
des3_speed_template, DES3_SPEED_VECTORS,
|
|
speed_template_24);
|
|
test_cipher_speed("ctr(des3_ede)", DECRYPT, sec,
|
|
des3_speed_template, DES3_SPEED_VECTORS,
|
|
speed_template_24);
|
|
break;
|
|
|
|
case 202:
|
|
test_cipher_speed("ecb(twofish)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_16_24_32);
|
|
test_cipher_speed("ecb(twofish)", DECRYPT, sec, NULL, 0,
|
|
speed_template_16_24_32);
|
|
test_cipher_speed("cbc(twofish)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_16_24_32);
|
|
test_cipher_speed("cbc(twofish)", DECRYPT, sec, NULL, 0,
|
|
speed_template_16_24_32);
|
|
test_cipher_speed("ctr(twofish)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_16_24_32);
|
|
test_cipher_speed("ctr(twofish)", DECRYPT, sec, NULL, 0,
|
|
speed_template_16_24_32);
|
|
test_cipher_speed("lrw(twofish)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_32_40_48);
|
|
test_cipher_speed("lrw(twofish)", DECRYPT, sec, NULL, 0,
|
|
speed_template_32_40_48);
|
|
test_cipher_speed("xts(twofish)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_32_48_64);
|
|
test_cipher_speed("xts(twofish)", DECRYPT, sec, NULL, 0,
|
|
speed_template_32_48_64);
|
|
break;
|
|
|
|
case 203:
|
|
test_cipher_speed("ecb(blowfish)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_8_32);
|
|
test_cipher_speed("ecb(blowfish)", DECRYPT, sec, NULL, 0,
|
|
speed_template_8_32);
|
|
test_cipher_speed("cbc(blowfish)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_8_32);
|
|
test_cipher_speed("cbc(blowfish)", DECRYPT, sec, NULL, 0,
|
|
speed_template_8_32);
|
|
test_cipher_speed("ctr(blowfish)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_8_32);
|
|
test_cipher_speed("ctr(blowfish)", DECRYPT, sec, NULL, 0,
|
|
speed_template_8_32);
|
|
break;
|
|
|
|
case 204:
|
|
test_cipher_speed("ecb(des)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_8);
|
|
test_cipher_speed("ecb(des)", DECRYPT, sec, NULL, 0,
|
|
speed_template_8);
|
|
test_cipher_speed("cbc(des)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_8);
|
|
test_cipher_speed("cbc(des)", DECRYPT, sec, NULL, 0,
|
|
speed_template_8);
|
|
break;
|
|
|
|
case 205:
|
|
test_cipher_speed("ecb(camellia)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_16_24_32);
|
|
test_cipher_speed("ecb(camellia)", DECRYPT, sec, NULL, 0,
|
|
speed_template_16_24_32);
|
|
test_cipher_speed("cbc(camellia)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_16_24_32);
|
|
test_cipher_speed("cbc(camellia)", DECRYPT, sec, NULL, 0,
|
|
speed_template_16_24_32);
|
|
test_cipher_speed("ctr(camellia)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_16_24_32);
|
|
test_cipher_speed("ctr(camellia)", DECRYPT, sec, NULL, 0,
|
|
speed_template_16_24_32);
|
|
test_cipher_speed("lrw(camellia)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_32_40_48);
|
|
test_cipher_speed("lrw(camellia)", DECRYPT, sec, NULL, 0,
|
|
speed_template_32_40_48);
|
|
test_cipher_speed("xts(camellia)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_32_48_64);
|
|
test_cipher_speed("xts(camellia)", DECRYPT, sec, NULL, 0,
|
|
speed_template_32_48_64);
|
|
break;
|
|
|
|
case 206:
|
|
test_cipher_speed("salsa20", ENCRYPT, sec, NULL, 0,
|
|
speed_template_16_32);
|
|
break;
|
|
|
|
case 207:
|
|
test_cipher_speed("ecb(serpent)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_16_32);
|
|
test_cipher_speed("ecb(serpent)", DECRYPT, sec, NULL, 0,
|
|
speed_template_16_32);
|
|
test_cipher_speed("cbc(serpent)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_16_32);
|
|
test_cipher_speed("cbc(serpent)", DECRYPT, sec, NULL, 0,
|
|
speed_template_16_32);
|
|
test_cipher_speed("ctr(serpent)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_16_32);
|
|
test_cipher_speed("ctr(serpent)", DECRYPT, sec, NULL, 0,
|
|
speed_template_16_32);
|
|
test_cipher_speed("lrw(serpent)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_32_48);
|
|
test_cipher_speed("lrw(serpent)", DECRYPT, sec, NULL, 0,
|
|
speed_template_32_48);
|
|
test_cipher_speed("xts(serpent)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_32_64);
|
|
test_cipher_speed("xts(serpent)", DECRYPT, sec, NULL, 0,
|
|
speed_template_32_64);
|
|
break;
|
|
|
|
case 208:
|
|
test_cipher_speed("ecb(arc4)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_8);
|
|
break;
|
|
|
|
case 209:
|
|
test_cipher_speed("ecb(cast5)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_8_16);
|
|
test_cipher_speed("ecb(cast5)", DECRYPT, sec, NULL, 0,
|
|
speed_template_8_16);
|
|
test_cipher_speed("cbc(cast5)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_8_16);
|
|
test_cipher_speed("cbc(cast5)", DECRYPT, sec, NULL, 0,
|
|
speed_template_8_16);
|
|
test_cipher_speed("ctr(cast5)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_8_16);
|
|
test_cipher_speed("ctr(cast5)", DECRYPT, sec, NULL, 0,
|
|
speed_template_8_16);
|
|
break;
|
|
|
|
case 210:
|
|
test_cipher_speed("ecb(cast6)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_16_32);
|
|
test_cipher_speed("ecb(cast6)", DECRYPT, sec, NULL, 0,
|
|
speed_template_16_32);
|
|
test_cipher_speed("cbc(cast6)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_16_32);
|
|
test_cipher_speed("cbc(cast6)", DECRYPT, sec, NULL, 0,
|
|
speed_template_16_32);
|
|
test_cipher_speed("ctr(cast6)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_16_32);
|
|
test_cipher_speed("ctr(cast6)", DECRYPT, sec, NULL, 0,
|
|
speed_template_16_32);
|
|
test_cipher_speed("lrw(cast6)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_32_48);
|
|
test_cipher_speed("lrw(cast6)", DECRYPT, sec, NULL, 0,
|
|
speed_template_32_48);
|
|
test_cipher_speed("xts(cast6)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_32_64);
|
|
test_cipher_speed("xts(cast6)", DECRYPT, sec, NULL, 0,
|
|
speed_template_32_64);
|
|
break;
|
|
|
|
case 211:
|
|
test_aead_speed("rfc4106(gcm(aes))", ENCRYPT, sec,
|
|
NULL, 0, 16, 16, aead_speed_template_20);
|
|
test_aead_speed("gcm(aes)", ENCRYPT, sec,
|
|
NULL, 0, 16, 8, speed_template_16_24_32);
|
|
test_aead_speed("rfc4106(gcm(aes))", DECRYPT, sec,
|
|
NULL, 0, 16, 16, aead_speed_template_20);
|
|
test_aead_speed("gcm(aes)", DECRYPT, sec,
|
|
NULL, 0, 16, 8, speed_template_16_24_32);
|
|
break;
|
|
|
|
case 212:
|
|
test_aead_speed("rfc4309(ccm(aes))", ENCRYPT, sec,
|
|
NULL, 0, 16, 16, aead_speed_template_19);
|
|
test_aead_speed("rfc4309(ccm(aes))", DECRYPT, sec,
|
|
NULL, 0, 16, 16, aead_speed_template_19);
|
|
break;
|
|
|
|
case 213:
|
|
test_aead_speed("rfc7539esp(chacha20,poly1305)", ENCRYPT, sec,
|
|
NULL, 0, 16, 8, aead_speed_template_36);
|
|
test_aead_speed("rfc7539esp(chacha20,poly1305)", DECRYPT, sec,
|
|
NULL, 0, 16, 8, aead_speed_template_36);
|
|
break;
|
|
|
|
case 214:
|
|
test_cipher_speed("chacha20", ENCRYPT, sec, NULL, 0,
|
|
speed_template_32);
|
|
break;
|
|
|
|
case 215:
|
|
test_mb_aead_speed("rfc4106(gcm(aes))", ENCRYPT, sec, NULL,
|
|
0, 16, 16, aead_speed_template_20, num_mb);
|
|
test_mb_aead_speed("gcm(aes)", ENCRYPT, sec, NULL, 0, 16, 8,
|
|
speed_template_16_24_32, num_mb);
|
|
test_mb_aead_speed("rfc4106(gcm(aes))", DECRYPT, sec, NULL,
|
|
0, 16, 16, aead_speed_template_20, num_mb);
|
|
test_mb_aead_speed("gcm(aes)", DECRYPT, sec, NULL, 0, 16, 8,
|
|
speed_template_16_24_32, num_mb);
|
|
break;
|
|
|
|
case 216:
|
|
test_mb_aead_speed("rfc4309(ccm(aes))", ENCRYPT, sec, NULL, 0,
|
|
16, 16, aead_speed_template_19, num_mb);
|
|
test_mb_aead_speed("rfc4309(ccm(aes))", DECRYPT, sec, NULL, 0,
|
|
16, 16, aead_speed_template_19, num_mb);
|
|
break;
|
|
|
|
case 217:
|
|
test_mb_aead_speed("rfc7539esp(chacha20,poly1305)", ENCRYPT,
|
|
sec, NULL, 0, 16, 8, aead_speed_template_36,
|
|
num_mb);
|
|
test_mb_aead_speed("rfc7539esp(chacha20,poly1305)", DECRYPT,
|
|
sec, NULL, 0, 16, 8, aead_speed_template_36,
|
|
num_mb);
|
|
break;
|
|
|
|
case 219:
|
|
test_cipher_speed("adiantum(xchacha12,aes)", ENCRYPT, sec, NULL,
|
|
0, speed_template_32);
|
|
test_cipher_speed("adiantum(xchacha12,aes)", DECRYPT, sec, NULL,
|
|
0, speed_template_32);
|
|
test_cipher_speed("adiantum(xchacha20,aes)", ENCRYPT, sec, NULL,
|
|
0, speed_template_32);
|
|
test_cipher_speed("adiantum(xchacha20,aes)", DECRYPT, sec, NULL,
|
|
0, speed_template_32);
|
|
break;
|
|
|
|
case 300:
|
|
if (alg) {
|
|
test_hash_speed(alg, sec, generic_hash_speed_template);
|
|
break;
|
|
}
|
|
/* fall through */
|
|
case 301:
|
|
test_hash_speed("md4", sec, generic_hash_speed_template);
|
|
if (mode > 300 && mode < 400) break;
|
|
/* fall through */
|
|
case 302:
|
|
test_hash_speed("md5", sec, generic_hash_speed_template);
|
|
if (mode > 300 && mode < 400) break;
|
|
/* fall through */
|
|
case 303:
|
|
test_hash_speed("sha1", sec, generic_hash_speed_template);
|
|
if (mode > 300 && mode < 400) break;
|
|
/* fall through */
|
|
case 304:
|
|
test_hash_speed("sha256", sec, generic_hash_speed_template);
|
|
if (mode > 300 && mode < 400) break;
|
|
/* fall through */
|
|
case 305:
|
|
test_hash_speed("sha384", sec, generic_hash_speed_template);
|
|
if (mode > 300 && mode < 400) break;
|
|
/* fall through */
|
|
case 306:
|
|
test_hash_speed("sha512", sec, generic_hash_speed_template);
|
|
if (mode > 300 && mode < 400) break;
|
|
/* fall through */
|
|
case 307:
|
|
test_hash_speed("wp256", sec, generic_hash_speed_template);
|
|
if (mode > 300 && mode < 400) break;
|
|
/* fall through */
|
|
case 308:
|
|
test_hash_speed("wp384", sec, generic_hash_speed_template);
|
|
if (mode > 300 && mode < 400) break;
|
|
/* fall through */
|
|
case 309:
|
|
test_hash_speed("wp512", sec, generic_hash_speed_template);
|
|
if (mode > 300 && mode < 400) break;
|
|
/* fall through */
|
|
case 310:
|
|
test_hash_speed("tgr128", sec, generic_hash_speed_template);
|
|
if (mode > 300 && mode < 400) break;
|
|
/* fall through */
|
|
case 311:
|
|
test_hash_speed("tgr160", sec, generic_hash_speed_template);
|
|
if (mode > 300 && mode < 400) break;
|
|
/* fall through */
|
|
case 312:
|
|
test_hash_speed("tgr192", sec, generic_hash_speed_template);
|
|
if (mode > 300 && mode < 400) break;
|
|
/* fall through */
|
|
case 313:
|
|
test_hash_speed("sha224", sec, generic_hash_speed_template);
|
|
if (mode > 300 && mode < 400) break;
|
|
/* fall through */
|
|
case 314:
|
|
test_hash_speed("rmd128", sec, generic_hash_speed_template);
|
|
if (mode > 300 && mode < 400) break;
|
|
/* fall through */
|
|
case 315:
|
|
test_hash_speed("rmd160", sec, generic_hash_speed_template);
|
|
if (mode > 300 && mode < 400) break;
|
|
/* fall through */
|
|
case 316:
|
|
test_hash_speed("rmd256", sec, generic_hash_speed_template);
|
|
if (mode > 300 && mode < 400) break;
|
|
/* fall through */
|
|
case 317:
|
|
test_hash_speed("rmd320", sec, generic_hash_speed_template);
|
|
if (mode > 300 && mode < 400) break;
|
|
/* fall through */
|
|
case 318:
|
|
test_hash_speed("ghash-generic", sec, hash_speed_template_16);
|
|
if (mode > 300 && mode < 400) break;
|
|
/* fall through */
|
|
case 319:
|
|
test_hash_speed("crc32c", sec, generic_hash_speed_template);
|
|
if (mode > 300 && mode < 400) break;
|
|
/* fall through */
|
|
case 320:
|
|
test_hash_speed("crct10dif", sec, generic_hash_speed_template);
|
|
if (mode > 300 && mode < 400) break;
|
|
/* fall through */
|
|
case 321:
|
|
test_hash_speed("poly1305", sec, poly1305_speed_template);
|
|
if (mode > 300 && mode < 400) break;
|
|
/* fall through */
|
|
case 322:
|
|
test_hash_speed("sha3-224", sec, generic_hash_speed_template);
|
|
if (mode > 300 && mode < 400) break;
|
|
/* fall through */
|
|
case 323:
|
|
test_hash_speed("sha3-256", sec, generic_hash_speed_template);
|
|
if (mode > 300 && mode < 400) break;
|
|
/* fall through */
|
|
case 324:
|
|
test_hash_speed("sha3-384", sec, generic_hash_speed_template);
|
|
if (mode > 300 && mode < 400) break;
|
|
/* fall through */
|
|
case 325:
|
|
test_hash_speed("sha3-512", sec, generic_hash_speed_template);
|
|
if (mode > 300 && mode < 400) break;
|
|
/* fall through */
|
|
case 326:
|
|
test_hash_speed("sm3", sec, generic_hash_speed_template);
|
|
if (mode > 300 && mode < 400) break;
|
|
/* fall through */
|
|
case 399:
|
|
break;
|
|
|
|
case 400:
|
|
if (alg) {
|
|
test_ahash_speed(alg, sec, generic_hash_speed_template);
|
|
break;
|
|
}
|
|
/* fall through */
|
|
case 401:
|
|
test_ahash_speed("md4", sec, generic_hash_speed_template);
|
|
if (mode > 400 && mode < 500) break;
|
|
/* fall through */
|
|
case 402:
|
|
test_ahash_speed("md5", sec, generic_hash_speed_template);
|
|
if (mode > 400 && mode < 500) break;
|
|
/* fall through */
|
|
case 403:
|
|
test_ahash_speed("sha1", sec, generic_hash_speed_template);
|
|
if (mode > 400 && mode < 500) break;
|
|
/* fall through */
|
|
case 404:
|
|
test_ahash_speed("sha256", sec, generic_hash_speed_template);
|
|
if (mode > 400 && mode < 500) break;
|
|
/* fall through */
|
|
case 405:
|
|
test_ahash_speed("sha384", sec, generic_hash_speed_template);
|
|
if (mode > 400 && mode < 500) break;
|
|
/* fall through */
|
|
case 406:
|
|
test_ahash_speed("sha512", sec, generic_hash_speed_template);
|
|
if (mode > 400 && mode < 500) break;
|
|
/* fall through */
|
|
case 407:
|
|
test_ahash_speed("wp256", sec, generic_hash_speed_template);
|
|
if (mode > 400 && mode < 500) break;
|
|
/* fall through */
|
|
case 408:
|
|
test_ahash_speed("wp384", sec, generic_hash_speed_template);
|
|
if (mode > 400 && mode < 500) break;
|
|
/* fall through */
|
|
case 409:
|
|
test_ahash_speed("wp512", sec, generic_hash_speed_template);
|
|
if (mode > 400 && mode < 500) break;
|
|
/* fall through */
|
|
case 410:
|
|
test_ahash_speed("tgr128", sec, generic_hash_speed_template);
|
|
if (mode > 400 && mode < 500) break;
|
|
/* fall through */
|
|
case 411:
|
|
test_ahash_speed("tgr160", sec, generic_hash_speed_template);
|
|
if (mode > 400 && mode < 500) break;
|
|
/* fall through */
|
|
case 412:
|
|
test_ahash_speed("tgr192", sec, generic_hash_speed_template);
|
|
if (mode > 400 && mode < 500) break;
|
|
/* fall through */
|
|
case 413:
|
|
test_ahash_speed("sha224", sec, generic_hash_speed_template);
|
|
if (mode > 400 && mode < 500) break;
|
|
/* fall through */
|
|
case 414:
|
|
test_ahash_speed("rmd128", sec, generic_hash_speed_template);
|
|
if (mode > 400 && mode < 500) break;
|
|
/* fall through */
|
|
case 415:
|
|
test_ahash_speed("rmd160", sec, generic_hash_speed_template);
|
|
if (mode > 400 && mode < 500) break;
|
|
/* fall through */
|
|
case 416:
|
|
test_ahash_speed("rmd256", sec, generic_hash_speed_template);
|
|
if (mode > 400 && mode < 500) break;
|
|
/* fall through */
|
|
case 417:
|
|
test_ahash_speed("rmd320", sec, generic_hash_speed_template);
|
|
if (mode > 400 && mode < 500) break;
|
|
/* fall through */
|
|
case 418:
|
|
test_ahash_speed("sha3-224", sec, generic_hash_speed_template);
|
|
if (mode > 400 && mode < 500) break;
|
|
/* fall through */
|
|
case 419:
|
|
test_ahash_speed("sha3-256", sec, generic_hash_speed_template);
|
|
if (mode > 400 && mode < 500) break;
|
|
/* fall through */
|
|
case 420:
|
|
test_ahash_speed("sha3-384", sec, generic_hash_speed_template);
|
|
if (mode > 400 && mode < 500) break;
|
|
/* fall through */
|
|
case 421:
|
|
test_ahash_speed("sha3-512", sec, generic_hash_speed_template);
|
|
if (mode > 400 && mode < 500) break;
|
|
/* fall through */
|
|
case 422:
|
|
test_mb_ahash_speed("sha1", sec, generic_hash_speed_template,
|
|
num_mb);
|
|
if (mode > 400 && mode < 500) break;
|
|
/* fall through */
|
|
case 423:
|
|
test_mb_ahash_speed("sha256", sec, generic_hash_speed_template,
|
|
num_mb);
|
|
if (mode > 400 && mode < 500) break;
|
|
/* fall through */
|
|
case 424:
|
|
test_mb_ahash_speed("sha512", sec, generic_hash_speed_template,
|
|
num_mb);
|
|
if (mode > 400 && mode < 500) break;
|
|
/* fall through */
|
|
case 425:
|
|
test_mb_ahash_speed("sm3", sec, generic_hash_speed_template,
|
|
num_mb);
|
|
if (mode > 400 && mode < 500) break;
|
|
/* fall through */
|
|
case 499:
|
|
break;
|
|
|
|
case 500:
|
|
test_acipher_speed("ecb(aes)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_16_24_32);
|
|
test_acipher_speed("ecb(aes)", DECRYPT, sec, NULL, 0,
|
|
speed_template_16_24_32);
|
|
test_acipher_speed("cbc(aes)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_16_24_32);
|
|
test_acipher_speed("cbc(aes)", DECRYPT, sec, NULL, 0,
|
|
speed_template_16_24_32);
|
|
test_acipher_speed("lrw(aes)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_32_40_48);
|
|
test_acipher_speed("lrw(aes)", DECRYPT, sec, NULL, 0,
|
|
speed_template_32_40_48);
|
|
test_acipher_speed("xts(aes)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_32_64);
|
|
test_acipher_speed("xts(aes)", DECRYPT, sec, NULL, 0,
|
|
speed_template_32_64);
|
|
test_acipher_speed("cts(cbc(aes))", ENCRYPT, sec, NULL, 0,
|
|
speed_template_16_24_32);
|
|
test_acipher_speed("cts(cbc(aes))", DECRYPT, sec, NULL, 0,
|
|
speed_template_16_24_32);
|
|
test_acipher_speed("ctr(aes)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_16_24_32);
|
|
test_acipher_speed("ctr(aes)", DECRYPT, sec, NULL, 0,
|
|
speed_template_16_24_32);
|
|
test_acipher_speed("cfb(aes)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_16_24_32);
|
|
test_acipher_speed("cfb(aes)", DECRYPT, sec, NULL, 0,
|
|
speed_template_16_24_32);
|
|
test_acipher_speed("ofb(aes)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_16_24_32);
|
|
test_acipher_speed("ofb(aes)", DECRYPT, sec, NULL, 0,
|
|
speed_template_16_24_32);
|
|
test_acipher_speed("rfc3686(ctr(aes))", ENCRYPT, sec, NULL, 0,
|
|
speed_template_20_28_36);
|
|
test_acipher_speed("rfc3686(ctr(aes))", DECRYPT, sec, NULL, 0,
|
|
speed_template_20_28_36);
|
|
break;
|
|
|
|
case 501:
|
|
test_acipher_speed("ecb(des3_ede)", ENCRYPT, sec,
|
|
des3_speed_template, DES3_SPEED_VECTORS,
|
|
speed_template_24);
|
|
test_acipher_speed("ecb(des3_ede)", DECRYPT, sec,
|
|
des3_speed_template, DES3_SPEED_VECTORS,
|
|
speed_template_24);
|
|
test_acipher_speed("cbc(des3_ede)", ENCRYPT, sec,
|
|
des3_speed_template, DES3_SPEED_VECTORS,
|
|
speed_template_24);
|
|
test_acipher_speed("cbc(des3_ede)", DECRYPT, sec,
|
|
des3_speed_template, DES3_SPEED_VECTORS,
|
|
speed_template_24);
|
|
test_acipher_speed("cfb(des3_ede)", ENCRYPT, sec,
|
|
des3_speed_template, DES3_SPEED_VECTORS,
|
|
speed_template_24);
|
|
test_acipher_speed("cfb(des3_ede)", DECRYPT, sec,
|
|
des3_speed_template, DES3_SPEED_VECTORS,
|
|
speed_template_24);
|
|
test_acipher_speed("ofb(des3_ede)", ENCRYPT, sec,
|
|
des3_speed_template, DES3_SPEED_VECTORS,
|
|
speed_template_24);
|
|
test_acipher_speed("ofb(des3_ede)", DECRYPT, sec,
|
|
des3_speed_template, DES3_SPEED_VECTORS,
|
|
speed_template_24);
|
|
break;
|
|
|
|
case 502:
|
|
test_acipher_speed("ecb(des)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_8);
|
|
test_acipher_speed("ecb(des)", DECRYPT, sec, NULL, 0,
|
|
speed_template_8);
|
|
test_acipher_speed("cbc(des)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_8);
|
|
test_acipher_speed("cbc(des)", DECRYPT, sec, NULL, 0,
|
|
speed_template_8);
|
|
test_acipher_speed("cfb(des)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_8);
|
|
test_acipher_speed("cfb(des)", DECRYPT, sec, NULL, 0,
|
|
speed_template_8);
|
|
test_acipher_speed("ofb(des)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_8);
|
|
test_acipher_speed("ofb(des)", DECRYPT, sec, NULL, 0,
|
|
speed_template_8);
|
|
break;
|
|
|
|
case 503:
|
|
test_acipher_speed("ecb(serpent)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_16_32);
|
|
test_acipher_speed("ecb(serpent)", DECRYPT, sec, NULL, 0,
|
|
speed_template_16_32);
|
|
test_acipher_speed("cbc(serpent)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_16_32);
|
|
test_acipher_speed("cbc(serpent)", DECRYPT, sec, NULL, 0,
|
|
speed_template_16_32);
|
|
test_acipher_speed("ctr(serpent)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_16_32);
|
|
test_acipher_speed("ctr(serpent)", DECRYPT, sec, NULL, 0,
|
|
speed_template_16_32);
|
|
test_acipher_speed("lrw(serpent)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_32_48);
|
|
test_acipher_speed("lrw(serpent)", DECRYPT, sec, NULL, 0,
|
|
speed_template_32_48);
|
|
test_acipher_speed("xts(serpent)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_32_64);
|
|
test_acipher_speed("xts(serpent)", DECRYPT, sec, NULL, 0,
|
|
speed_template_32_64);
|
|
break;
|
|
|
|
case 504:
|
|
test_acipher_speed("ecb(twofish)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_16_24_32);
|
|
test_acipher_speed("ecb(twofish)", DECRYPT, sec, NULL, 0,
|
|
speed_template_16_24_32);
|
|
test_acipher_speed("cbc(twofish)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_16_24_32);
|
|
test_acipher_speed("cbc(twofish)", DECRYPT, sec, NULL, 0,
|
|
speed_template_16_24_32);
|
|
test_acipher_speed("ctr(twofish)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_16_24_32);
|
|
test_acipher_speed("ctr(twofish)", DECRYPT, sec, NULL, 0,
|
|
speed_template_16_24_32);
|
|
test_acipher_speed("lrw(twofish)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_32_40_48);
|
|
test_acipher_speed("lrw(twofish)", DECRYPT, sec, NULL, 0,
|
|
speed_template_32_40_48);
|
|
test_acipher_speed("xts(twofish)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_32_48_64);
|
|
test_acipher_speed("xts(twofish)", DECRYPT, sec, NULL, 0,
|
|
speed_template_32_48_64);
|
|
break;
|
|
|
|
case 505:
|
|
test_acipher_speed("ecb(arc4)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_8);
|
|
break;
|
|
|
|
case 506:
|
|
test_acipher_speed("ecb(cast5)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_8_16);
|
|
test_acipher_speed("ecb(cast5)", DECRYPT, sec, NULL, 0,
|
|
speed_template_8_16);
|
|
test_acipher_speed("cbc(cast5)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_8_16);
|
|
test_acipher_speed("cbc(cast5)", DECRYPT, sec, NULL, 0,
|
|
speed_template_8_16);
|
|
test_acipher_speed("ctr(cast5)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_8_16);
|
|
test_acipher_speed("ctr(cast5)", DECRYPT, sec, NULL, 0,
|
|
speed_template_8_16);
|
|
break;
|
|
|
|
case 507:
|
|
test_acipher_speed("ecb(cast6)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_16_32);
|
|
test_acipher_speed("ecb(cast6)", DECRYPT, sec, NULL, 0,
|
|
speed_template_16_32);
|
|
test_acipher_speed("cbc(cast6)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_16_32);
|
|
test_acipher_speed("cbc(cast6)", DECRYPT, sec, NULL, 0,
|
|
speed_template_16_32);
|
|
test_acipher_speed("ctr(cast6)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_16_32);
|
|
test_acipher_speed("ctr(cast6)", DECRYPT, sec, NULL, 0,
|
|
speed_template_16_32);
|
|
test_acipher_speed("lrw(cast6)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_32_48);
|
|
test_acipher_speed("lrw(cast6)", DECRYPT, sec, NULL, 0,
|
|
speed_template_32_48);
|
|
test_acipher_speed("xts(cast6)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_32_64);
|
|
test_acipher_speed("xts(cast6)", DECRYPT, sec, NULL, 0,
|
|
speed_template_32_64);
|
|
break;
|
|
|
|
case 508:
|
|
test_acipher_speed("ecb(camellia)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_16_32);
|
|
test_acipher_speed("ecb(camellia)", DECRYPT, sec, NULL, 0,
|
|
speed_template_16_32);
|
|
test_acipher_speed("cbc(camellia)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_16_32);
|
|
test_acipher_speed("cbc(camellia)", DECRYPT, sec, NULL, 0,
|
|
speed_template_16_32);
|
|
test_acipher_speed("ctr(camellia)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_16_32);
|
|
test_acipher_speed("ctr(camellia)", DECRYPT, sec, NULL, 0,
|
|
speed_template_16_32);
|
|
test_acipher_speed("lrw(camellia)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_32_48);
|
|
test_acipher_speed("lrw(camellia)", DECRYPT, sec, NULL, 0,
|
|
speed_template_32_48);
|
|
test_acipher_speed("xts(camellia)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_32_64);
|
|
test_acipher_speed("xts(camellia)", DECRYPT, sec, NULL, 0,
|
|
speed_template_32_64);
|
|
break;
|
|
|
|
case 509:
|
|
test_acipher_speed("ecb(blowfish)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_8_32);
|
|
test_acipher_speed("ecb(blowfish)", DECRYPT, sec, NULL, 0,
|
|
speed_template_8_32);
|
|
test_acipher_speed("cbc(blowfish)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_8_32);
|
|
test_acipher_speed("cbc(blowfish)", DECRYPT, sec, NULL, 0,
|
|
speed_template_8_32);
|
|
test_acipher_speed("ctr(blowfish)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_8_32);
|
|
test_acipher_speed("ctr(blowfish)", DECRYPT, sec, NULL, 0,
|
|
speed_template_8_32);
|
|
break;
|
|
|
|
case 600:
|
|
test_mb_skcipher_speed("ecb(aes)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_16_24_32, num_mb);
|
|
test_mb_skcipher_speed("ecb(aes)", DECRYPT, sec, NULL, 0,
|
|
speed_template_16_24_32, num_mb);
|
|
test_mb_skcipher_speed("cbc(aes)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_16_24_32, num_mb);
|
|
test_mb_skcipher_speed("cbc(aes)", DECRYPT, sec, NULL, 0,
|
|
speed_template_16_24_32, num_mb);
|
|
test_mb_skcipher_speed("lrw(aes)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_32_40_48, num_mb);
|
|
test_mb_skcipher_speed("lrw(aes)", DECRYPT, sec, NULL, 0,
|
|
speed_template_32_40_48, num_mb);
|
|
test_mb_skcipher_speed("xts(aes)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_32_64, num_mb);
|
|
test_mb_skcipher_speed("xts(aes)", DECRYPT, sec, NULL, 0,
|
|
speed_template_32_64, num_mb);
|
|
test_mb_skcipher_speed("cts(cbc(aes))", ENCRYPT, sec, NULL, 0,
|
|
speed_template_16_24_32, num_mb);
|
|
test_mb_skcipher_speed("cts(cbc(aes))", DECRYPT, sec, NULL, 0,
|
|
speed_template_16_24_32, num_mb);
|
|
test_mb_skcipher_speed("ctr(aes)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_16_24_32, num_mb);
|
|
test_mb_skcipher_speed("ctr(aes)", DECRYPT, sec, NULL, 0,
|
|
speed_template_16_24_32, num_mb);
|
|
test_mb_skcipher_speed("cfb(aes)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_16_24_32, num_mb);
|
|
test_mb_skcipher_speed("cfb(aes)", DECRYPT, sec, NULL, 0,
|
|
speed_template_16_24_32, num_mb);
|
|
test_mb_skcipher_speed("ofb(aes)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_16_24_32, num_mb);
|
|
test_mb_skcipher_speed("ofb(aes)", DECRYPT, sec, NULL, 0,
|
|
speed_template_16_24_32, num_mb);
|
|
test_mb_skcipher_speed("rfc3686(ctr(aes))", ENCRYPT, sec, NULL,
|
|
0, speed_template_20_28_36, num_mb);
|
|
test_mb_skcipher_speed("rfc3686(ctr(aes))", DECRYPT, sec, NULL,
|
|
0, speed_template_20_28_36, num_mb);
|
|
break;
|
|
|
|
case 601:
|
|
test_mb_skcipher_speed("ecb(des3_ede)", ENCRYPT, sec,
|
|
des3_speed_template, DES3_SPEED_VECTORS,
|
|
speed_template_24, num_mb);
|
|
test_mb_skcipher_speed("ecb(des3_ede)", DECRYPT, sec,
|
|
des3_speed_template, DES3_SPEED_VECTORS,
|
|
speed_template_24, num_mb);
|
|
test_mb_skcipher_speed("cbc(des3_ede)", ENCRYPT, sec,
|
|
des3_speed_template, DES3_SPEED_VECTORS,
|
|
speed_template_24, num_mb);
|
|
test_mb_skcipher_speed("cbc(des3_ede)", DECRYPT, sec,
|
|
des3_speed_template, DES3_SPEED_VECTORS,
|
|
speed_template_24, num_mb);
|
|
test_mb_skcipher_speed("cfb(des3_ede)", ENCRYPT, sec,
|
|
des3_speed_template, DES3_SPEED_VECTORS,
|
|
speed_template_24, num_mb);
|
|
test_mb_skcipher_speed("cfb(des3_ede)", DECRYPT, sec,
|
|
des3_speed_template, DES3_SPEED_VECTORS,
|
|
speed_template_24, num_mb);
|
|
test_mb_skcipher_speed("ofb(des3_ede)", ENCRYPT, sec,
|
|
des3_speed_template, DES3_SPEED_VECTORS,
|
|
speed_template_24, num_mb);
|
|
test_mb_skcipher_speed("ofb(des3_ede)", DECRYPT, sec,
|
|
des3_speed_template, DES3_SPEED_VECTORS,
|
|
speed_template_24, num_mb);
|
|
break;
|
|
|
|
case 602:
|
|
test_mb_skcipher_speed("ecb(des)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_8, num_mb);
|
|
test_mb_skcipher_speed("ecb(des)", DECRYPT, sec, NULL, 0,
|
|
speed_template_8, num_mb);
|
|
test_mb_skcipher_speed("cbc(des)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_8, num_mb);
|
|
test_mb_skcipher_speed("cbc(des)", DECRYPT, sec, NULL, 0,
|
|
speed_template_8, num_mb);
|
|
test_mb_skcipher_speed("cfb(des)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_8, num_mb);
|
|
test_mb_skcipher_speed("cfb(des)", DECRYPT, sec, NULL, 0,
|
|
speed_template_8, num_mb);
|
|
test_mb_skcipher_speed("ofb(des)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_8, num_mb);
|
|
test_mb_skcipher_speed("ofb(des)", DECRYPT, sec, NULL, 0,
|
|
speed_template_8, num_mb);
|
|
break;
|
|
|
|
case 603:
|
|
test_mb_skcipher_speed("ecb(serpent)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_16_32, num_mb);
|
|
test_mb_skcipher_speed("ecb(serpent)", DECRYPT, sec, NULL, 0,
|
|
speed_template_16_32, num_mb);
|
|
test_mb_skcipher_speed("cbc(serpent)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_16_32, num_mb);
|
|
test_mb_skcipher_speed("cbc(serpent)", DECRYPT, sec, NULL, 0,
|
|
speed_template_16_32, num_mb);
|
|
test_mb_skcipher_speed("ctr(serpent)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_16_32, num_mb);
|
|
test_mb_skcipher_speed("ctr(serpent)", DECRYPT, sec, NULL, 0,
|
|
speed_template_16_32, num_mb);
|
|
test_mb_skcipher_speed("lrw(serpent)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_32_48, num_mb);
|
|
test_mb_skcipher_speed("lrw(serpent)", DECRYPT, sec, NULL, 0,
|
|
speed_template_32_48, num_mb);
|
|
test_mb_skcipher_speed("xts(serpent)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_32_64, num_mb);
|
|
test_mb_skcipher_speed("xts(serpent)", DECRYPT, sec, NULL, 0,
|
|
speed_template_32_64, num_mb);
|
|
break;
|
|
|
|
case 604:
|
|
test_mb_skcipher_speed("ecb(twofish)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_16_24_32, num_mb);
|
|
test_mb_skcipher_speed("ecb(twofish)", DECRYPT, sec, NULL, 0,
|
|
speed_template_16_24_32, num_mb);
|
|
test_mb_skcipher_speed("cbc(twofish)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_16_24_32, num_mb);
|
|
test_mb_skcipher_speed("cbc(twofish)", DECRYPT, sec, NULL, 0,
|
|
speed_template_16_24_32, num_mb);
|
|
test_mb_skcipher_speed("ctr(twofish)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_16_24_32, num_mb);
|
|
test_mb_skcipher_speed("ctr(twofish)", DECRYPT, sec, NULL, 0,
|
|
speed_template_16_24_32, num_mb);
|
|
test_mb_skcipher_speed("lrw(twofish)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_32_40_48, num_mb);
|
|
test_mb_skcipher_speed("lrw(twofish)", DECRYPT, sec, NULL, 0,
|
|
speed_template_32_40_48, num_mb);
|
|
test_mb_skcipher_speed("xts(twofish)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_32_48_64, num_mb);
|
|
test_mb_skcipher_speed("xts(twofish)", DECRYPT, sec, NULL, 0,
|
|
speed_template_32_48_64, num_mb);
|
|
break;
|
|
|
|
case 605:
|
|
test_mb_skcipher_speed("ecb(arc4)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_8, num_mb);
|
|
break;
|
|
|
|
case 606:
|
|
test_mb_skcipher_speed("ecb(cast5)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_8_16, num_mb);
|
|
test_mb_skcipher_speed("ecb(cast5)", DECRYPT, sec, NULL, 0,
|
|
speed_template_8_16, num_mb);
|
|
test_mb_skcipher_speed("cbc(cast5)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_8_16, num_mb);
|
|
test_mb_skcipher_speed("cbc(cast5)", DECRYPT, sec, NULL, 0,
|
|
speed_template_8_16, num_mb);
|
|
test_mb_skcipher_speed("ctr(cast5)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_8_16, num_mb);
|
|
test_mb_skcipher_speed("ctr(cast5)", DECRYPT, sec, NULL, 0,
|
|
speed_template_8_16, num_mb);
|
|
break;
|
|
|
|
case 607:
|
|
test_mb_skcipher_speed("ecb(cast6)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_16_32, num_mb);
|
|
test_mb_skcipher_speed("ecb(cast6)", DECRYPT, sec, NULL, 0,
|
|
speed_template_16_32, num_mb);
|
|
test_mb_skcipher_speed("cbc(cast6)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_16_32, num_mb);
|
|
test_mb_skcipher_speed("cbc(cast6)", DECRYPT, sec, NULL, 0,
|
|
speed_template_16_32, num_mb);
|
|
test_mb_skcipher_speed("ctr(cast6)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_16_32, num_mb);
|
|
test_mb_skcipher_speed("ctr(cast6)", DECRYPT, sec, NULL, 0,
|
|
speed_template_16_32, num_mb);
|
|
test_mb_skcipher_speed("lrw(cast6)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_32_48, num_mb);
|
|
test_mb_skcipher_speed("lrw(cast6)", DECRYPT, sec, NULL, 0,
|
|
speed_template_32_48, num_mb);
|
|
test_mb_skcipher_speed("xts(cast6)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_32_64, num_mb);
|
|
test_mb_skcipher_speed("xts(cast6)", DECRYPT, sec, NULL, 0,
|
|
speed_template_32_64, num_mb);
|
|
break;
|
|
|
|
case 608:
|
|
test_mb_skcipher_speed("ecb(camellia)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_16_32, num_mb);
|
|
test_mb_skcipher_speed("ecb(camellia)", DECRYPT, sec, NULL, 0,
|
|
speed_template_16_32, num_mb);
|
|
test_mb_skcipher_speed("cbc(camellia)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_16_32, num_mb);
|
|
test_mb_skcipher_speed("cbc(camellia)", DECRYPT, sec, NULL, 0,
|
|
speed_template_16_32, num_mb);
|
|
test_mb_skcipher_speed("ctr(camellia)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_16_32, num_mb);
|
|
test_mb_skcipher_speed("ctr(camellia)", DECRYPT, sec, NULL, 0,
|
|
speed_template_16_32, num_mb);
|
|
test_mb_skcipher_speed("lrw(camellia)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_32_48, num_mb);
|
|
test_mb_skcipher_speed("lrw(camellia)", DECRYPT, sec, NULL, 0,
|
|
speed_template_32_48, num_mb);
|
|
test_mb_skcipher_speed("xts(camellia)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_32_64, num_mb);
|
|
test_mb_skcipher_speed("xts(camellia)", DECRYPT, sec, NULL, 0,
|
|
speed_template_32_64, num_mb);
|
|
break;
|
|
|
|
case 609:
|
|
test_mb_skcipher_speed("ecb(blowfish)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_8_32, num_mb);
|
|
test_mb_skcipher_speed("ecb(blowfish)", DECRYPT, sec, NULL, 0,
|
|
speed_template_8_32, num_mb);
|
|
test_mb_skcipher_speed("cbc(blowfish)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_8_32, num_mb);
|
|
test_mb_skcipher_speed("cbc(blowfish)", DECRYPT, sec, NULL, 0,
|
|
speed_template_8_32, num_mb);
|
|
test_mb_skcipher_speed("ctr(blowfish)", ENCRYPT, sec, NULL, 0,
|
|
speed_template_8_32, num_mb);
|
|
test_mb_skcipher_speed("ctr(blowfish)", DECRYPT, sec, NULL, 0,
|
|
speed_template_8_32, num_mb);
|
|
break;
|
|
|
|
case 1000:
|
|
test_available();
|
|
break;
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int __init tcrypt_mod_init(void)
|
|
{
|
|
int err = -ENOMEM;
|
|
int i;
|
|
|
|
for (i = 0; i < TVMEMSIZE; i++) {
|
|
tvmem[i] = (void *)__get_free_page(GFP_KERNEL);
|
|
if (!tvmem[i])
|
|
goto err_free_tv;
|
|
}
|
|
|
|
err = do_test(alg, type, mask, mode, num_mb);
|
|
|
|
if (err) {
|
|
printk(KERN_ERR "tcrypt: one or more tests failed!\n");
|
|
goto err_free_tv;
|
|
} else {
|
|
pr_debug("all tests passed\n");
|
|
}
|
|
|
|
/* We intentionaly return -EAGAIN to prevent keeping the module,
|
|
* unless we're running in fips mode. It does all its work from
|
|
* init() and doesn't offer any runtime functionality, but in
|
|
* the fips case, checking for a successful load is helpful.
|
|
* => we don't need it in the memory, do we?
|
|
* -- mludvig
|
|
*/
|
|
if (!fips_enabled)
|
|
err = -EAGAIN;
|
|
|
|
err_free_tv:
|
|
for (i = 0; i < TVMEMSIZE && tvmem[i]; i++)
|
|
free_page((unsigned long)tvmem[i]);
|
|
|
|
return err;
|
|
}
|
|
|
|
/*
|
|
* If an init function is provided, an exit function must also be provided
|
|
* to allow module unload.
|
|
*/
|
|
static void __exit tcrypt_mod_fini(void) { }
|
|
|
|
module_init(tcrypt_mod_init);
|
|
module_exit(tcrypt_mod_fini);
|
|
|
|
module_param(alg, charp, 0);
|
|
module_param(type, uint, 0);
|
|
module_param(mask, uint, 0);
|
|
module_param(mode, int, 0);
|
|
module_param(sec, uint, 0);
|
|
MODULE_PARM_DESC(sec, "Length in seconds of speed tests "
|
|
"(defaults to zero which uses CPU cycles instead)");
|
|
module_param(num_mb, uint, 0000);
|
|
MODULE_PARM_DESC(num_mb, "Number of concurrent requests to be used in mb speed tests (defaults to 8)");
|
|
|
|
MODULE_LICENSE("GPL");
|
|
MODULE_DESCRIPTION("Quick & dirty crypto testing module");
|
|
MODULE_AUTHOR("James Morris <jmorris@intercode.com.au>");
|