mirror of
https://github.com/openssl/openssl.git
synced 2025-01-12 13:36:28 +08:00
cc82b09cbd
Unroll AES-CTR loops to a maximum 12 blocks for ARM Neoverse V1 and V2, to fully utilize their AES pipeline resources. Improvement on ARM Neoverse V1. Package Size(Bytes) 16 32 64 128 256 1024 Improvement(%) 3.93 -0.45 11.30 4.31 12.48 37.66 Package Size(Bytes) 1500 8192 16384 61440 65536 Improvement(%) 37.16 38.90 39.89 40.55 40.41 Change-Id: Ifb8fad9af22476259b9ba75132bc3d8010a7fdbd Reviewed-by: Tom Cosgrove <tom.cosgrove@arm.com> Reviewed-by: Tomas Mraz <tomas@openssl.org> (Merged from https://github.com/openssl/openssl/pull/22733)
113 lines
3.8 KiB
C++
113 lines
3.8 KiB
C++
/*
|
|
* Copyright 2019-2023 The OpenSSL Project Authors. All Rights Reserved.
|
|
*
|
|
* Licensed under the Apache License 2.0 (the "License"). You may not use
|
|
* this file except in compliance with the License. You can obtain a copy
|
|
* in the file LICENSE in the source distribution or at
|
|
* https://www.openssl.org/source/license.html
|
|
*/
|
|
|
|
/*
|
|
* Crypto extension support for AES GCM.
|
|
* This file is included by cipher_aes_gcm_hw.c
|
|
*/
|
|
|
|
size_t armv8_aes_gcm_encrypt(const unsigned char *in, unsigned char *out, size_t len,
|
|
const void *key, unsigned char ivec[16], u64 *Xi)
|
|
{
|
|
size_t align_bytes = 0;
|
|
align_bytes = len - len % 16;
|
|
|
|
AES_KEY *aes_key = (AES_KEY *)key;
|
|
|
|
switch(aes_key->rounds) {
|
|
case 10:
|
|
if (IS_CPU_SUPPORT_UNROLL8_EOR3()) {
|
|
unroll8_eor3_aes_gcm_enc_128_kernel(in, align_bytes * 8, out, (uint64_t *)Xi, ivec, key);
|
|
} else {
|
|
aes_gcm_enc_128_kernel(in, align_bytes * 8, out, (uint64_t *)Xi, ivec, key);
|
|
}
|
|
break;
|
|
case 12:
|
|
if (IS_CPU_SUPPORT_UNROLL8_EOR3()) {
|
|
unroll8_eor3_aes_gcm_enc_192_kernel(in, align_bytes * 8, out, (uint64_t *)Xi, ivec, key);
|
|
} else {
|
|
aes_gcm_enc_192_kernel(in, align_bytes * 8, out, (uint64_t *)Xi, ivec, key);
|
|
}
|
|
break;
|
|
case 14:
|
|
if (IS_CPU_SUPPORT_UNROLL8_EOR3()) {
|
|
unroll8_eor3_aes_gcm_enc_256_kernel(in, align_bytes * 8, out, (uint64_t *)Xi, ivec, key);
|
|
} else {
|
|
aes_gcm_enc_256_kernel(in, align_bytes * 8, out, (uint64_t *)Xi, ivec, key);
|
|
}
|
|
break;
|
|
}
|
|
return align_bytes;
|
|
}
|
|
|
|
size_t armv8_aes_gcm_decrypt(const unsigned char *in, unsigned char *out, size_t len,
|
|
const void *key, unsigned char ivec[16], u64 *Xi)
|
|
{
|
|
size_t align_bytes = 0;
|
|
align_bytes = len - len % 16;
|
|
|
|
AES_KEY *aes_key = (AES_KEY *)key;
|
|
|
|
switch(aes_key->rounds) {
|
|
case 10:
|
|
if (IS_CPU_SUPPORT_UNROLL8_EOR3()) {
|
|
unroll8_eor3_aes_gcm_dec_128_kernel(in, align_bytes * 8, out, (uint64_t *)Xi, ivec, key);
|
|
} else {
|
|
aes_gcm_dec_128_kernel(in, align_bytes * 8, out, (uint64_t *)Xi, ivec, key);
|
|
}
|
|
break;
|
|
case 12:
|
|
if (IS_CPU_SUPPORT_UNROLL8_EOR3()) {
|
|
unroll8_eor3_aes_gcm_dec_192_kernel(in, align_bytes * 8, out, (uint64_t *)Xi, ivec, key);
|
|
} else {
|
|
aes_gcm_dec_192_kernel(in, align_bytes * 8, out, (uint64_t *)Xi, ivec, key);
|
|
}
|
|
break;
|
|
case 14:
|
|
if (IS_CPU_SUPPORT_UNROLL8_EOR3()) {
|
|
unroll8_eor3_aes_gcm_dec_256_kernel(in, align_bytes * 8, out, (uint64_t *)Xi, ivec, key);
|
|
} else {
|
|
aes_gcm_dec_256_kernel(in, align_bytes * 8, out, (uint64_t *)Xi, ivec, key);
|
|
}
|
|
break;
|
|
}
|
|
return align_bytes;
|
|
}
|
|
|
|
static int armv8_aes_gcm_initkey(PROV_GCM_CTX *ctx, const unsigned char *key,
|
|
size_t keylen)
|
|
{
|
|
PROV_AES_GCM_CTX *actx = (PROV_AES_GCM_CTX *)ctx;
|
|
AES_KEY *ks = &actx->ks.ks;
|
|
|
|
if (AES_UNROLL12_EOR3_CAPABLE) {
|
|
GCM_HW_SET_KEY_CTR_FN(ks, aes_v8_set_encrypt_key, aes_v8_encrypt,
|
|
aes_v8_ctr32_encrypt_blocks_unroll12_eor3);
|
|
} else {
|
|
GCM_HW_SET_KEY_CTR_FN(ks, aes_v8_set_encrypt_key, aes_v8_encrypt,
|
|
aes_v8_ctr32_encrypt_blocks);
|
|
}
|
|
return 1;
|
|
}
|
|
|
|
|
|
static const PROV_GCM_HW armv8_aes_gcm = {
|
|
armv8_aes_gcm_initkey,
|
|
ossl_gcm_setiv,
|
|
ossl_gcm_aad_update,
|
|
generic_aes_gcm_cipher_update,
|
|
ossl_gcm_cipher_final,
|
|
ossl_gcm_one_shot
|
|
};
|
|
|
|
const PROV_GCM_HW *ossl_prov_aes_hw_gcm(size_t keybits)
|
|
{
|
|
return AES_PMULL_CAPABLE ? &armv8_aes_gcm : &aes_gcm;
|
|
}
|