2019-08-26 15:05:08 +08:00
|
|
|
/*
|
2020-04-23 20:55:52 +08:00
|
|
|
* Copyright 2019-2020 The OpenSSL Project Authors. All Rights Reserved.
|
2019-08-26 15:05:08 +08:00
|
|
|
*
|
|
|
|
* Licensed under the Apache License 2.0 (the "License"). You may not use
|
|
|
|
* this file except in compliance with the License. You can obtain a copy
|
|
|
|
* in the file LICENSE in the source distribution or at
|
|
|
|
* https://www.openssl.org/source/license.html
|
|
|
|
*/
|
|
|
|
|
|
|
|
/* Dispatch functions for AES GCM mode */
|
|
|
|
|
2019-12-06 01:09:49 +08:00
|
|
|
/*
|
|
|
|
* This file uses the low level AES functions (which are deprecated for
|
|
|
|
* non-internal use) in order to implement provider AES ciphers.
|
|
|
|
*/
|
|
|
|
#include "internal/deprecated.h"
|
|
|
|
|
2019-12-11 20:43:24 +08:00
|
|
|
#include "cipher_aes_gcm.h"
|
2019-08-26 15:05:08 +08:00
|
|
|
|
2020-01-08 21:58:34 +08:00
|
|
|
static int aes_gcm_initkey(PROV_GCM_CTX *ctx, const unsigned char *key,
|
2019-08-26 15:05:08 +08:00
|
|
|
size_t keylen)
|
|
|
|
{
|
|
|
|
PROV_AES_GCM_CTX *actx = (PROV_AES_GCM_CTX *)ctx;
|
|
|
|
AES_KEY *ks = &actx->ks.ks;
|
|
|
|
|
|
|
|
# ifdef HWAES_CAPABLE
|
|
|
|
if (HWAES_CAPABLE) {
|
|
|
|
# ifdef HWAES_ctr32_encrypt_blocks
|
|
|
|
GCM_HW_SET_KEY_CTR_FN(ks, HWAES_set_encrypt_key, HWAES_encrypt,
|
|
|
|
HWAES_ctr32_encrypt_blocks);
|
|
|
|
# else
|
|
|
|
GCM_HW_SET_KEY_CTR_FN(ks, HWAES_set_encrypt_key, HWAES_encrypt, NULL);
|
|
|
|
# endif /* HWAES_ctr32_encrypt_blocks */
|
|
|
|
} else
|
|
|
|
# endif /* HWAES_CAPABLE */
|
|
|
|
|
|
|
|
# ifdef BSAES_CAPABLE
|
|
|
|
if (BSAES_CAPABLE) {
|
|
|
|
GCM_HW_SET_KEY_CTR_FN(ks, AES_set_encrypt_key, AES_encrypt,
|
|
|
|
bsaes_ctr32_encrypt_blocks);
|
|
|
|
} else
|
|
|
|
# endif /* BSAES_CAPABLE */
|
|
|
|
|
|
|
|
# ifdef VPAES_CAPABLE
|
|
|
|
if (VPAES_CAPABLE) {
|
|
|
|
GCM_HW_SET_KEY_CTR_FN(ks, vpaes_set_encrypt_key, vpaes_encrypt, NULL);
|
|
|
|
} else
|
|
|
|
# endif /* VPAES_CAPABLE */
|
|
|
|
|
|
|
|
{
|
|
|
|
# ifdef AES_CTR_ASM
|
|
|
|
GCM_HW_SET_KEY_CTR_FN(ks, AES_set_encrypt_key, AES_encrypt,
|
|
|
|
AES_ctr32_encrypt);
|
|
|
|
# else
|
|
|
|
GCM_HW_SET_KEY_CTR_FN(ks, AES_set_encrypt_key, AES_encrypt, NULL);
|
|
|
|
# endif /* AES_CTR_ASM */
|
|
|
|
}
|
|
|
|
ctx->key_set = 1;
|
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
|
2020-01-08 21:58:34 +08:00
|
|
|
static int generic_aes_gcm_cipher_update(PROV_GCM_CTX *ctx, const unsigned char *in,
|
|
|
|
size_t len, unsigned char *out)
|
|
|
|
{
|
|
|
|
if (ctx->enc) {
|
|
|
|
if (ctx->ctr != NULL) {
|
|
|
|
#if defined(AES_GCM_ASM)
|
|
|
|
size_t bulk = 0;
|
|
|
|
|
|
|
|
if (len >= AES_GCM_ENC_BYTES && AES_GCM_ASM(ctx)) {
|
|
|
|
size_t res = (16 - ctx->gcm.mres) % 16;
|
|
|
|
|
|
|
|
if (CRYPTO_gcm128_encrypt(&ctx->gcm, in, out, res))
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
bulk = AES_gcm_encrypt(in + res, out + res, len - res,
|
|
|
|
ctx->gcm.key,
|
|
|
|
ctx->gcm.Yi.c, ctx->gcm.Xi.u);
|
|
|
|
|
|
|
|
ctx->gcm.len.u[1] += bulk;
|
|
|
|
bulk += res;
|
|
|
|
}
|
|
|
|
if (CRYPTO_gcm128_encrypt_ctr32(&ctx->gcm, in + bulk, out + bulk,
|
|
|
|
len - bulk, ctx->ctr))
|
|
|
|
return 0;
|
|
|
|
#else
|
|
|
|
if (CRYPTO_gcm128_encrypt_ctr32(&ctx->gcm, in, out, len, ctx->ctr))
|
|
|
|
return 0;
|
|
|
|
#endif /* AES_GCM_ASM */
|
|
|
|
} else {
|
|
|
|
if (CRYPTO_gcm128_encrypt(&ctx->gcm, in, out, len))
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
if (ctx->ctr != NULL) {
|
|
|
|
#if defined(AES_GCM_ASM)
|
|
|
|
size_t bulk = 0;
|
|
|
|
|
|
|
|
if (len >= AES_GCM_DEC_BYTES && AES_GCM_ASM(ctx)) {
|
|
|
|
size_t res = (16 - ctx->gcm.mres) % 16;
|
|
|
|
|
|
|
|
if (CRYPTO_gcm128_decrypt(&ctx->gcm, in, out, res))
|
|
|
|
return -1;
|
|
|
|
|
|
|
|
bulk = AES_gcm_decrypt(in + res, out + res, len - res,
|
|
|
|
ctx->gcm.key,
|
|
|
|
ctx->gcm.Yi.c, ctx->gcm.Xi.u);
|
|
|
|
|
|
|
|
ctx->gcm.len.u[1] += bulk;
|
|
|
|
bulk += res;
|
|
|
|
}
|
|
|
|
if (CRYPTO_gcm128_decrypt_ctr32(&ctx->gcm, in + bulk, out + bulk,
|
|
|
|
len - bulk, ctx->ctr))
|
|
|
|
return 0;
|
|
|
|
#else
|
|
|
|
if (CRYPTO_gcm128_decrypt_ctr32(&ctx->gcm, in, out, len, ctx->ctr))
|
|
|
|
return 0;
|
|
|
|
#endif /* AES_GCM_ASM */
|
|
|
|
} else {
|
|
|
|
if (CRYPTO_gcm128_decrypt(&ctx->gcm, in, out, len))
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
|
2019-08-26 15:05:08 +08:00
|
|
|
static const PROV_GCM_HW aes_gcm = {
|
2020-01-08 21:58:34 +08:00
|
|
|
aes_gcm_initkey,
|
2019-08-26 15:05:08 +08:00
|
|
|
gcm_setiv,
|
|
|
|
gcm_aad_update,
|
2020-01-08 21:58:34 +08:00
|
|
|
generic_aes_gcm_cipher_update,
|
2019-08-26 15:05:08 +08:00
|
|
|
gcm_cipher_final,
|
|
|
|
gcm_one_shot
|
|
|
|
};
|
|
|
|
|
|
|
|
#if defined(S390X_aes_128_CAPABLE)
|
|
|
|
# include "cipher_aes_gcm_hw_s390x.inc"
|
|
|
|
#elif defined(AESNI_CAPABLE)
|
|
|
|
# include "cipher_aes_gcm_hw_aesni.inc"
|
|
|
|
#elif defined(SPARC_AES_CAPABLE)
|
|
|
|
# include "cipher_aes_gcm_hw_t4.inc"
|
2019-05-31 18:15:10 +08:00
|
|
|
#elif defined(AES_PMULL_CAPABLE) && defined(AES_GCM_ASM)
|
|
|
|
# include "cipher_aes_gcm_hw_armv8.inc"
|
2019-08-26 15:05:08 +08:00
|
|
|
#else
|
|
|
|
const PROV_GCM_HW *PROV_AES_HW_gcm(size_t keybits)
|
|
|
|
{
|
|
|
|
return &aes_gcm;
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
|