Merge git://git.kernel.org/pub/scm/linux/kernel/git/herbert/crypto-2.6
* git://git.kernel.org/pub/scm/linux/kernel/git/herbert/crypto-2.6: crypto: gcm - fix another complete call in complete fuction crypto: padlock-aes - Use the correct mask when checking whether copying is required
This commit is contained in:
commit
52c7b3f45d
2 changed files with 75 additions and 36 deletions
107
crypto/gcm.c
107
crypto/gcm.c
|
@ -40,7 +40,7 @@ struct crypto_rfc4106_ctx {
|
|||
struct crypto_gcm_ghash_ctx {
|
||||
unsigned int cryptlen;
|
||||
struct scatterlist *src;
|
||||
crypto_completion_t complete;
|
||||
void (*complete)(struct aead_request *req, int err);
|
||||
};
|
||||
|
||||
struct crypto_gcm_req_priv_ctx {
|
||||
|
@ -267,23 +267,26 @@ static int gcm_hash_final(struct aead_request *req,
|
|||
return crypto_ahash_final(ahreq);
|
||||
}
|
||||
|
||||
static void gcm_hash_final_done(struct crypto_async_request *areq,
|
||||
int err)
|
||||
static void __gcm_hash_final_done(struct aead_request *req, int err)
|
||||
{
|
||||
struct aead_request *req = areq->data;
|
||||
struct crypto_gcm_req_priv_ctx *pctx = crypto_gcm_reqctx(req);
|
||||
struct crypto_gcm_ghash_ctx *gctx = &pctx->ghash_ctx;
|
||||
|
||||
if (!err)
|
||||
crypto_xor(pctx->auth_tag, pctx->iauth_tag, 16);
|
||||
|
||||
gctx->complete(areq, err);
|
||||
gctx->complete(req, err);
|
||||
}
|
||||
|
||||
static void gcm_hash_len_done(struct crypto_async_request *areq,
|
||||
int err)
|
||||
static void gcm_hash_final_done(struct crypto_async_request *areq, int err)
|
||||
{
|
||||
struct aead_request *req = areq->data;
|
||||
|
||||
__gcm_hash_final_done(req, err);
|
||||
}
|
||||
|
||||
static void __gcm_hash_len_done(struct aead_request *req, int err)
|
||||
{
|
||||
struct crypto_gcm_req_priv_ctx *pctx = crypto_gcm_reqctx(req);
|
||||
|
||||
if (!err) {
|
||||
|
@ -292,13 +295,18 @@ static void gcm_hash_len_done(struct crypto_async_request *areq,
|
|||
return;
|
||||
}
|
||||
|
||||
gcm_hash_final_done(areq, err);
|
||||
__gcm_hash_final_done(req, err);
|
||||
}
|
||||
|
||||
static void gcm_hash_crypt_remain_done(struct crypto_async_request *areq,
|
||||
int err)
|
||||
static void gcm_hash_len_done(struct crypto_async_request *areq, int err)
|
||||
{
|
||||
struct aead_request *req = areq->data;
|
||||
|
||||
__gcm_hash_len_done(req, err);
|
||||
}
|
||||
|
||||
static void __gcm_hash_crypt_remain_done(struct aead_request *req, int err)
|
||||
{
|
||||
struct crypto_gcm_req_priv_ctx *pctx = crypto_gcm_reqctx(req);
|
||||
|
||||
if (!err) {
|
||||
|
@ -307,13 +315,19 @@ static void gcm_hash_crypt_remain_done(struct crypto_async_request *areq,
|
|||
return;
|
||||
}
|
||||
|
||||
gcm_hash_len_done(areq, err);
|
||||
__gcm_hash_len_done(req, err);
|
||||
}
|
||||
|
||||
static void gcm_hash_crypt_done(struct crypto_async_request *areq,
|
||||
int err)
|
||||
static void gcm_hash_crypt_remain_done(struct crypto_async_request *areq,
|
||||
int err)
|
||||
{
|
||||
struct aead_request *req = areq->data;
|
||||
|
||||
__gcm_hash_crypt_remain_done(req, err);
|
||||
}
|
||||
|
||||
static void __gcm_hash_crypt_done(struct aead_request *req, int err)
|
||||
{
|
||||
struct crypto_gcm_req_priv_ctx *pctx = crypto_gcm_reqctx(req);
|
||||
struct crypto_gcm_ghash_ctx *gctx = &pctx->ghash_ctx;
|
||||
unsigned int remain;
|
||||
|
@ -327,13 +341,18 @@ static void gcm_hash_crypt_done(struct crypto_async_request *areq,
|
|||
return;
|
||||
}
|
||||
|
||||
gcm_hash_crypt_remain_done(areq, err);
|
||||
__gcm_hash_crypt_remain_done(req, err);
|
||||
}
|
||||
|
||||
static void gcm_hash_assoc_remain_done(struct crypto_async_request *areq,
|
||||
int err)
|
||||
static void gcm_hash_crypt_done(struct crypto_async_request *areq, int err)
|
||||
{
|
||||
struct aead_request *req = areq->data;
|
||||
|
||||
__gcm_hash_crypt_done(req, err);
|
||||
}
|
||||
|
||||
static void __gcm_hash_assoc_remain_done(struct aead_request *req, int err)
|
||||
{
|
||||
struct crypto_gcm_req_priv_ctx *pctx = crypto_gcm_reqctx(req);
|
||||
struct crypto_gcm_ghash_ctx *gctx = &pctx->ghash_ctx;
|
||||
crypto_completion_t complete;
|
||||
|
@ -350,15 +369,21 @@ static void gcm_hash_assoc_remain_done(struct crypto_async_request *areq,
|
|||
}
|
||||
|
||||
if (remain)
|
||||
gcm_hash_crypt_done(areq, err);
|
||||
__gcm_hash_crypt_done(req, err);
|
||||
else
|
||||
gcm_hash_crypt_remain_done(areq, err);
|
||||
__gcm_hash_crypt_remain_done(req, err);
|
||||
}
|
||||
|
||||
static void gcm_hash_assoc_done(struct crypto_async_request *areq,
|
||||
int err)
|
||||
static void gcm_hash_assoc_remain_done(struct crypto_async_request *areq,
|
||||
int err)
|
||||
{
|
||||
struct aead_request *req = areq->data;
|
||||
|
||||
__gcm_hash_assoc_remain_done(req, err);
|
||||
}
|
||||
|
||||
static void __gcm_hash_assoc_done(struct aead_request *req, int err)
|
||||
{
|
||||
struct crypto_gcm_req_priv_ctx *pctx = crypto_gcm_reqctx(req);
|
||||
unsigned int remain;
|
||||
|
||||
|
@ -371,13 +396,18 @@ static void gcm_hash_assoc_done(struct crypto_async_request *areq,
|
|||
return;
|
||||
}
|
||||
|
||||
gcm_hash_assoc_remain_done(areq, err);
|
||||
__gcm_hash_assoc_remain_done(req, err);
|
||||
}
|
||||
|
||||
static void gcm_hash_init_done(struct crypto_async_request *areq,
|
||||
int err)
|
||||
static void gcm_hash_assoc_done(struct crypto_async_request *areq, int err)
|
||||
{
|
||||
struct aead_request *req = areq->data;
|
||||
|
||||
__gcm_hash_assoc_done(req, err);
|
||||
}
|
||||
|
||||
static void __gcm_hash_init_done(struct aead_request *req, int err)
|
||||
{
|
||||
struct crypto_gcm_req_priv_ctx *pctx = crypto_gcm_reqctx(req);
|
||||
crypto_completion_t complete;
|
||||
unsigned int remain = 0;
|
||||
|
@ -393,9 +423,16 @@ static void gcm_hash_init_done(struct crypto_async_request *areq,
|
|||
}
|
||||
|
||||
if (remain)
|
||||
gcm_hash_assoc_done(areq, err);
|
||||
__gcm_hash_assoc_done(req, err);
|
||||
else
|
||||
gcm_hash_assoc_remain_done(areq, err);
|
||||
__gcm_hash_assoc_remain_done(req, err);
|
||||
}
|
||||
|
||||
static void gcm_hash_init_done(struct crypto_async_request *areq, int err)
|
||||
{
|
||||
struct aead_request *req = areq->data;
|
||||
|
||||
__gcm_hash_init_done(req, err);
|
||||
}
|
||||
|
||||
static int gcm_hash(struct aead_request *req,
|
||||
|
@ -457,10 +494,8 @@ static void gcm_enc_copy_hash(struct aead_request *req,
|
|||
crypto_aead_authsize(aead), 1);
|
||||
}
|
||||
|
||||
static void gcm_enc_hash_done(struct crypto_async_request *areq,
|
||||
int err)
|
||||
static void gcm_enc_hash_done(struct aead_request *req, int err)
|
||||
{
|
||||
struct aead_request *req = areq->data;
|
||||
struct crypto_gcm_req_priv_ctx *pctx = crypto_gcm_reqctx(req);
|
||||
|
||||
if (!err)
|
||||
|
@ -469,8 +504,7 @@ static void gcm_enc_hash_done(struct crypto_async_request *areq,
|
|||
aead_request_complete(req, err);
|
||||
}
|
||||
|
||||
static void gcm_encrypt_done(struct crypto_async_request *areq,
|
||||
int err)
|
||||
static void gcm_encrypt_done(struct crypto_async_request *areq, int err)
|
||||
{
|
||||
struct aead_request *req = areq->data;
|
||||
struct crypto_gcm_req_priv_ctx *pctx = crypto_gcm_reqctx(req);
|
||||
|
@ -479,9 +513,13 @@ static void gcm_encrypt_done(struct crypto_async_request *areq,
|
|||
err = gcm_hash(req, pctx);
|
||||
if (err == -EINPROGRESS || err == -EBUSY)
|
||||
return;
|
||||
else if (!err) {
|
||||
crypto_xor(pctx->auth_tag, pctx->iauth_tag, 16);
|
||||
gcm_enc_copy_hash(req, pctx);
|
||||
}
|
||||
}
|
||||
|
||||
gcm_enc_hash_done(areq, err);
|
||||
aead_request_complete(req, err);
|
||||
}
|
||||
|
||||
static int crypto_gcm_encrypt(struct aead_request *req)
|
||||
|
@ -538,9 +576,8 @@ static void gcm_decrypt_done(struct crypto_async_request *areq, int err)
|
|||
aead_request_complete(req, err);
|
||||
}
|
||||
|
||||
static void gcm_dec_hash_done(struct crypto_async_request *areq, int err)
|
||||
static void gcm_dec_hash_done(struct aead_request *req, int err)
|
||||
{
|
||||
struct aead_request *req = areq->data;
|
||||
struct crypto_gcm_req_priv_ctx *pctx = crypto_gcm_reqctx(req);
|
||||
struct ablkcipher_request *abreq = &pctx->u.abreq;
|
||||
struct crypto_gcm_ghash_ctx *gctx = &pctx->ghash_ctx;
|
||||
|
@ -552,9 +589,11 @@ static void gcm_dec_hash_done(struct crypto_async_request *areq, int err)
|
|||
err = crypto_ablkcipher_decrypt(abreq);
|
||||
if (err == -EINPROGRESS || err == -EBUSY)
|
||||
return;
|
||||
else if (!err)
|
||||
err = crypto_gcm_verify(req, pctx);
|
||||
}
|
||||
|
||||
gcm_decrypt_done(areq, err);
|
||||
aead_request_complete(req, err);
|
||||
}
|
||||
|
||||
static int crypto_gcm_decrypt(struct aead_request *req)
|
||||
|
|
|
@ -236,7 +236,7 @@ static inline void ecb_crypt(const u8 *in, u8 *out, u32 *key,
|
|||
/* Padlock in ECB mode fetches at least ecb_fetch_bytes of data.
|
||||
* We could avoid some copying here but it's probably not worth it.
|
||||
*/
|
||||
if (unlikely(((unsigned long)in & PAGE_SIZE) + ecb_fetch_bytes > PAGE_SIZE)) {
|
||||
if (unlikely(((unsigned long)in & ~PAGE_MASK) + ecb_fetch_bytes > PAGE_SIZE)) {
|
||||
ecb_crypt_copy(in, out, key, cword, count);
|
||||
return;
|
||||
}
|
||||
|
@ -248,7 +248,7 @@ static inline u8 *cbc_crypt(const u8 *in, u8 *out, u32 *key,
|
|||
u8 *iv, struct cword *cword, int count)
|
||||
{
|
||||
/* Padlock in CBC mode fetches at least cbc_fetch_bytes of data. */
|
||||
if (unlikely(((unsigned long)in & PAGE_SIZE) + cbc_fetch_bytes > PAGE_SIZE))
|
||||
if (unlikely(((unsigned long)in & ~PAGE_MASK) + cbc_fetch_bytes > PAGE_SIZE))
|
||||
return cbc_crypt_copy(in, out, key, iv, cword, count);
|
||||
|
||||
return rep_xcrypt_cbc(in, out, key, iv, cword, count);
|
||||
|
|
Loading…
Reference in a new issue