Skip to content

Commit

Permalink
crypto: tcrypt - Add speed tests for SHA multibuffer algorithms
Browse files Browse the repository at this point in the history
The existing test suite to calculate the speed of the SHA algorithms
assumes serial (single buffer)) computation of data. With the SHA
multibuffer algorithms, we work on 8 lanes of data in parallel. Hence,
the need to introduce a new test suite to calculate the speed for these
algorithms.

Signed-off-by: Megha Dey <megha.dey@linux.intel.com>
Signed-off-by: Herbert Xu <herbert@gondor.apana.org.au>
  • Loading branch information
Megha Dey authored and Herbert Xu committed Jun 27, 2016
1 parent 9925324 commit 087bcd2
Showing 1 changed file with 118 additions and 0 deletions.
118 changes: 118 additions & 0 deletions crypto/tcrypt.c
Original file line number Diff line number Diff line change
Expand Up @@ -578,6 +578,117 @@ static inline int do_one_ahash_op(struct ahash_request *req, int ret)
return ret;
}

char ptext[4096];
struct scatterlist sg[8][8];
char result[8][64];
struct ahash_request *req[8];
struct tcrypt_result tresult[8];
char *xbuf[8][XBUFSIZE];
cycles_t start[8], end[8], mid;

static void test_mb_ahash_speed(const char *algo, unsigned int sec,
struct hash_speed *speed)
{
unsigned int i, j, k;
void *hash_buff;
int ret = -ENOMEM;
struct crypto_ahash *tfm;

tfm = crypto_alloc_ahash(algo, 0, 0);
if (IS_ERR(tfm)) {
pr_err("failed to load transform for %s: %ld\n",
algo, PTR_ERR(tfm));
return;
}
for (i = 0; i < 8; ++i) {
if (testmgr_alloc_buf(xbuf[i]))
goto out_nobuf;

init_completion(&tresult[i].completion);

req[i] = ahash_request_alloc(tfm, GFP_KERNEL);
if (!req[i]) {
printk(KERN_ERR "alg: hash: Failed to allocate "
"request for %s\n", algo);
goto out_noreq;
}
ahash_request_set_callback(req[i], CRYPTO_TFM_REQ_MAY_BACKLOG,
tcrypt_complete, &tresult[i]);

hash_buff = xbuf[i][0];
memcpy(hash_buff, ptext, 4096);
}

j = 0;

printk(KERN_INFO "\ntesting speed of %s (%s)\n", algo,
get_driver_name(crypto_ahash, tfm));

for (i = 0; speed[i].blen != 0; i++) {
if (speed[i].blen > TVMEMSIZE * PAGE_SIZE) {
printk(KERN_ERR
"template (%u) too big for tvmem (%lu)\n",
speed[i].blen, TVMEMSIZE * PAGE_SIZE);
goto out;
}

if (speed[i].klen)
crypto_ahash_setkey(tfm, tvmem[0], speed[i].klen);

for (k = 0; k < 8; ++k) {
sg_init_one(&sg[k][0], (void *) xbuf[k][0],
speed[i].blen);
ahash_request_set_crypt(req[k], sg[k],
result[k], speed[i].blen);
}

printk(KERN_INFO "test%3u "
"(%5u byte blocks,%5u bytes per update,%4u updates): ",
i, speed[i].blen, speed[i].plen,
speed[i].blen / speed[i].plen);

for (k = 0; k < 8; ++k) {
start[k] = get_cycles();
ret = crypto_ahash_digest(req[k]);
if (ret == -EBUSY || ret == -EINPROGRESS)
continue;
if (ret) {
printk(KERN_ERR
"alg (%s) something wrong, ret = %d ...\n",
algo, ret);
goto out;
}
}
mid = get_cycles();

for (k = 0; k < 8; ++k) {
struct tcrypt_result *tr = &tresult[k];

ret = wait_for_completion_interruptible
(&tr->completion);
if (ret)
printk(KERN_ERR
"alg(%s): hash: digest failed\n", algo);
end[k] = get_cycles();
}

printk("\nBlock: %lld cycles (%lld cycles/byte), %d bytes\n",
(s64) (end[7]-start[0])/1,
(s64) (end[7]-start[0])/(8*speed[i].blen),
8*speed[i].blen);
}
ret = 0;

out:
for (k = 0; k < 8; ++k)
ahash_request_free(req[k]);
out_noreq:
for (k = 0; k < 8; ++k)
testmgr_free_buf(xbuf[k]);
out_nobuf:
return;
}

static int test_ahash_jiffies_digest(struct ahash_request *req, int blen,
char *out, int secs)
{
Expand Down Expand Up @@ -1820,6 +1931,13 @@ static int do_test(const char *alg, u32 type, u32 mask, int m)
test_ahash_speed("sha3-512", sec, generic_hash_speed_template);
if (mode > 400 && mode < 500) break;

case 422:
test_mb_ahash_speed("sha1", sec, generic_hash_speed_template);
if (mode > 400 && mode < 500) break;

case 423:
test_mb_ahash_speed("sha256", sec, generic_hash_speed_template);
if (mode > 400 && mode < 500) break;

case 499:
break;
Expand Down

0 comments on commit 087bcd2

Please sign in to comment.