mirror of
https://github.com/openssl/openssl.git
synced 2024-12-21 06:09:35 +08:00
2788b56f0c
Signed-off-by: Xu Yizhou <xuyizhou1@huawei.com> Reviewed-by: Hugo Landau <hlandau@openssl.org> Reviewed-by: Tomas Mraz <tomas@openssl.org> (Merged from https://github.com/openssl/openssl/pull/19619)
237 lines
7.7 KiB
C
237 lines
7.7 KiB
C
/*
|
|
* Copyright 2010-2022 The OpenSSL Project Authors. All Rights Reserved.
|
|
*
|
|
* Licensed under the Apache License 2.0 (the "License"). You may not use
|
|
* this file except in compliance with the License. You can obtain a copy
|
|
* in the file LICENSE in the source distribution or at
|
|
* https://www.openssl.org/source/license.html
|
|
*/
|
|
|
|
/* This header can move into provider when legacy support is removed */
|
|
#include <openssl/modes.h>
|
|
|
|
#if (defined(_WIN32) || defined(_WIN64)) && !defined(__MINGW32__)
|
|
typedef __int64 i64;
|
|
typedef unsigned __int64 u64;
|
|
# define U64(C) C##UI64
|
|
#elif defined(__arch64__)
|
|
typedef long i64;
|
|
typedef unsigned long u64;
|
|
# define U64(C) C##UL
|
|
#else
|
|
typedef long long i64;
|
|
typedef unsigned long long u64;
|
|
# define U64(C) C##ULL
|
|
#endif
|
|
|
|
typedef unsigned int u32;
|
|
typedef unsigned char u8;
|
|
|
|
#define STRICT_ALIGNMENT 1
|
|
#ifndef PEDANTIC
|
|
# if defined(__i386) || defined(__i386__) || \
|
|
defined(__x86_64) || defined(__x86_64__) || \
|
|
defined(_M_IX86) || defined(_M_AMD64) || defined(_M_X64) || \
|
|
defined(__aarch64__) || \
|
|
defined(__s390__) || defined(__s390x__)
|
|
# undef STRICT_ALIGNMENT
|
|
# endif
|
|
#endif
|
|
|
|
#if !defined(PEDANTIC) && !defined(OPENSSL_NO_ASM) && !defined(OPENSSL_NO_INLINE_ASM)
|
|
# if defined(__GNUC__) && __GNUC__>=2
|
|
# if defined(__x86_64) || defined(__x86_64__)
|
|
# define BSWAP8(x) ({ u64 ret_=(x); \
|
|
asm ("bswapq %0" \
|
|
: "+r"(ret_)); ret_; })
|
|
# define BSWAP4(x) ({ u32 ret_=(x); \
|
|
asm ("bswapl %0" \
|
|
: "+r"(ret_)); ret_; })
|
|
# elif (defined(__i386) || defined(__i386__)) && !defined(I386_ONLY)
|
|
# define BSWAP8(x) ({ u32 lo_=(u64)(x)>>32,hi_=(x); \
|
|
asm ("bswapl %0; bswapl %1" \
|
|
: "+r"(hi_),"+r"(lo_)); \
|
|
(u64)hi_<<32|lo_; })
|
|
# define BSWAP4(x) ({ u32 ret_=(x); \
|
|
asm ("bswapl %0" \
|
|
: "+r"(ret_)); ret_; })
|
|
# elif defined(__aarch64__)
|
|
# if defined(__BYTE_ORDER__) && defined(__ORDER_LITTLE_ENDIAN__) && \
|
|
__BYTE_ORDER__==__ORDER_LITTLE_ENDIAN__
|
|
# define BSWAP8(x) ({ u64 ret_; \
|
|
asm ("rev %0,%1" \
|
|
: "=r"(ret_) : "r"(x)); ret_; })
|
|
# define BSWAP4(x) ({ u32 ret_; \
|
|
asm ("rev %w0,%w1" \
|
|
: "=r"(ret_) : "r"(x)); ret_; })
|
|
# endif
|
|
# elif (defined(__arm__) || defined(__arm)) && !defined(STRICT_ALIGNMENT)
|
|
# define BSWAP8(x) ({ u32 lo_=(u64)(x)>>32,hi_=(x); \
|
|
asm ("rev %0,%0; rev %1,%1" \
|
|
: "+r"(hi_),"+r"(lo_)); \
|
|
(u64)hi_<<32|lo_; })
|
|
# define BSWAP4(x) ({ u32 ret_; \
|
|
asm ("rev %0,%1" \
|
|
: "=r"(ret_) : "r"((u32)(x))); \
|
|
ret_; })
|
|
# elif (defined(__riscv_zbb) || defined(__riscv_zbkb)) && __riscv_xlen == 64
|
|
# define BSWAP8(x) ({ u64 ret_=(x); \
|
|
asm ("rev8 %0,%0" \
|
|
: "+r"(ret_)); ret_; })
|
|
# define BSWAP4(x) ({ u32 ret_=(x); \
|
|
asm ("rev8 %0,%0; srli %0,%0,32"\
|
|
: "+&r"(ret_)); ret_; })
|
|
# endif
|
|
# elif defined(_MSC_VER)
|
|
# if _MSC_VER>=1300
|
|
# include <stdlib.h>
|
|
# pragma intrinsic(_byteswap_uint64,_byteswap_ulong)
|
|
# define BSWAP8(x) _byteswap_uint64((u64)(x))
|
|
# define BSWAP4(x) _byteswap_ulong((u32)(x))
|
|
# elif defined(_M_IX86)
|
|
__inline u32 _bswap4(u32 val)
|
|
{
|
|
_asm mov eax, val _asm bswap eax}
|
|
# define BSWAP4(x) _bswap4(x)
|
|
# endif
|
|
# endif
|
|
#endif
|
|
#if defined(BSWAP4) && !defined(STRICT_ALIGNMENT)
|
|
# define GETU32(p) BSWAP4(*(const u32 *)(p))
|
|
# define PUTU32(p,v) *(u32 *)(p) = BSWAP4(v)
|
|
#else
|
|
# define GETU32(p) ((u32)(p)[0]<<24|(u32)(p)[1]<<16|(u32)(p)[2]<<8|(u32)(p)[3])
|
|
# define PUTU32(p,v) ((p)[0]=(u8)((v)>>24),(p)[1]=(u8)((v)>>16),(p)[2]=(u8)((v)>>8),(p)[3]=(u8)(v))
|
|
#endif
|
|
/*- GCM definitions */ typedef struct {
|
|
u64 hi, lo;
|
|
} u128;
|
|
|
|
typedef void (*gcm_init_fn)(u128 Htable[16], const u64 H[2]);
|
|
typedef void (*gcm_ghash_fn)(u64 Xi[2], const u128 Htable[16], const u8 *inp, size_t len);
|
|
typedef void (*gcm_gmult_fn)(u64 Xi[2], const u128 Htable[16]);
|
|
struct gcm_funcs_st {
|
|
gcm_init_fn ginit;
|
|
gcm_ghash_fn ghash;
|
|
gcm_gmult_fn gmult;
|
|
};
|
|
|
|
struct gcm128_context {
|
|
/* Following 6 names follow names in GCM specification */
|
|
union {
|
|
u64 u[2];
|
|
u32 d[4];
|
|
u8 c[16];
|
|
size_t t[16 / sizeof(size_t)];
|
|
} Yi, EKi, EK0, len, Xi, H;
|
|
/*
|
|
* Relative position of Yi, EKi, EK0, len, Xi, H and pre-computed Htable is
|
|
* used in some assembler modules, i.e. don't change the order!
|
|
*/
|
|
u128 Htable[16];
|
|
struct gcm_funcs_st funcs;
|
|
unsigned int mres, ares;
|
|
block128_f block;
|
|
void *key;
|
|
#if !defined(OPENSSL_SMALL_FOOTPRINT)
|
|
unsigned char Xn[48];
|
|
#endif
|
|
};
|
|
|
|
/* GHASH functions */
|
|
void ossl_gcm_init_4bit(u128 Htable[16], const u64 H[2]);
|
|
void ossl_gcm_ghash_4bit(u64 Xi[2], const u128 Htable[16],
|
|
const u8 *inp, size_t len);
|
|
void ossl_gcm_gmult_4bit(u64 Xi[2], const u128 Htable[16]);
|
|
|
|
/*
|
|
* The maximum permitted number of cipher blocks per data unit in XTS mode.
|
|
* Reference IEEE Std 1619-2018.
|
|
*/
|
|
#define XTS_MAX_BLOCKS_PER_DATA_UNIT (1<<20)
|
|
|
|
struct xts128_context {
|
|
void *key1, *key2;
|
|
block128_f block1, block2;
|
|
};
|
|
|
|
/* XTS mode for SM4 algorithm specified by GB/T 17964-2021 */
|
|
int ossl_crypto_xts128gb_encrypt(const XTS128_CONTEXT *ctx,
|
|
const unsigned char iv[16],
|
|
const unsigned char *inp, unsigned char *out,
|
|
size_t len, int enc);
|
|
|
|
struct ccm128_context {
|
|
union {
|
|
u64 u[2];
|
|
u8 c[16];
|
|
} nonce, cmac;
|
|
u64 blocks;
|
|
block128_f block;
|
|
void *key;
|
|
};
|
|
|
|
#ifndef OPENSSL_NO_OCB
|
|
|
|
typedef union {
|
|
u64 a[2];
|
|
unsigned char c[16];
|
|
} OCB_BLOCK;
|
|
# define ocb_block16_xor(in1,in2,out) \
|
|
( (out)->a[0]=(in1)->a[0]^(in2)->a[0], \
|
|
(out)->a[1]=(in1)->a[1]^(in2)->a[1] )
|
|
# if STRICT_ALIGNMENT
|
|
# define ocb_block16_xor_misaligned(in1,in2,out) \
|
|
ocb_block_xor((in1)->c,(in2)->c,16,(out)->c)
|
|
# else
|
|
# define ocb_block16_xor_misaligned ocb_block16_xor
|
|
# endif
|
|
|
|
struct ocb128_context {
|
|
/* Need both encrypt and decrypt key schedules for decryption */
|
|
block128_f encrypt;
|
|
block128_f decrypt;
|
|
void *keyenc;
|
|
void *keydec;
|
|
ocb128_f stream; /* direction dependent */
|
|
/* Key dependent variables. Can be reused if key remains the same */
|
|
size_t l_index;
|
|
size_t max_l_index;
|
|
OCB_BLOCK l_star;
|
|
OCB_BLOCK l_dollar;
|
|
OCB_BLOCK *l;
|
|
/* Must be reset for each session */
|
|
struct {
|
|
u64 blocks_hashed;
|
|
u64 blocks_processed;
|
|
OCB_BLOCK offset_aad;
|
|
OCB_BLOCK sum;
|
|
OCB_BLOCK offset;
|
|
OCB_BLOCK checksum;
|
|
} sess;
|
|
};
|
|
#endif /* OPENSSL_NO_OCB */
|
|
|
|
#ifndef OPENSSL_NO_SIV
|
|
|
|
#define SIV_LEN 16
|
|
|
|
typedef union siv_block_u {
|
|
uint64_t word[SIV_LEN/sizeof(uint64_t)];
|
|
unsigned char byte[SIV_LEN];
|
|
} SIV_BLOCK;
|
|
|
|
struct siv128_context {
|
|
/* d stores intermediate results of S2V; it corresponds to D from the
|
|
pseudocode in section 2.4 of RFC 5297. */
|
|
SIV_BLOCK d;
|
|
SIV_BLOCK tag;
|
|
EVP_CIPHER_CTX *cipher_ctx;
|
|
EVP_MAC *mac;
|
|
EVP_MAC_CTX *mac_ctx_init;
|
|
int final_ret;
|
|
int crypto_ok;
|
|
};
|
|
|
|
#endif /* OPENSSL_NO_SIV */
|