]> git.ipfire.org Git - thirdparty/openssl.git/blob - providers/implementations/ciphers/cipher_aes_gcm_hw_armv8.inc
Update copyright year
[thirdparty/openssl.git] / providers / implementations / ciphers / cipher_aes_gcm_hw_armv8.inc
1 /*
2 * Copyright 2019-2022 The OpenSSL Project Authors. All Rights Reserved.
3 *
4 * Licensed under the Apache License 2.0 (the "License"). You may not use
5 * this file except in compliance with the License. You can obtain a copy
6 * in the file LICENSE in the source distribution or at
7 * https://www.openssl.org/source/license.html
8 */
9
10 /*
11 * Crypto extension support for AES GCM.
12 * This file is included by cipher_aes_gcm_hw.c
13 */
14
15 size_t armv8_aes_gcm_encrypt(const unsigned char *in, unsigned char *out, size_t len,
16 const void *key, unsigned char ivec[16], u64 *Xi)
17 {
18 size_t align_bytes = 0;
19 align_bytes = len - len % 16;
20
21 AES_KEY *aes_key = (AES_KEY *)key;
22
23 switch(aes_key->rounds) {
24 case 10:
25 if (IS_CPU_SUPPORT_UNROLL8_EOR3()) {
26 unroll8_eor3_aes_gcm_enc_128_kernel(in, align_bytes * 8, out, (uint64_t *)Xi, ivec, key);
27 } else {
28 aes_gcm_enc_128_kernel(in, align_bytes * 8, out, (uint64_t *)Xi, ivec, key);
29 }
30 break;
31 case 12:
32 if (IS_CPU_SUPPORT_UNROLL8_EOR3()) {
33 unroll8_eor3_aes_gcm_enc_192_kernel(in, align_bytes * 8, out, (uint64_t *)Xi, ivec, key);
34 } else {
35 aes_gcm_enc_192_kernel(in, align_bytes * 8, out, (uint64_t *)Xi, ivec, key);
36 }
37 break;
38 case 14:
39 if (IS_CPU_SUPPORT_UNROLL8_EOR3()) {
40 unroll8_eor3_aes_gcm_enc_256_kernel(in, align_bytes * 8, out, (uint64_t *)Xi, ivec, key);
41 } else {
42 aes_gcm_enc_256_kernel(in, align_bytes * 8, out, (uint64_t *)Xi, ivec, key);
43 }
44 break;
45 }
46 return align_bytes;
47 }
48
49 size_t armv8_aes_gcm_decrypt(const unsigned char *in, unsigned char *out, size_t len,
50 const void *key, unsigned char ivec[16], u64 *Xi)
51 {
52 size_t align_bytes = 0;
53 align_bytes = len - len % 16;
54
55 AES_KEY *aes_key = (AES_KEY *)key;
56
57 switch(aes_key->rounds) {
58 case 10:
59 if (IS_CPU_SUPPORT_UNROLL8_EOR3()) {
60 unroll8_eor3_aes_gcm_dec_128_kernel(in, align_bytes * 8, out, (uint64_t *)Xi, ivec, key);
61 } else {
62 aes_gcm_dec_128_kernel(in, align_bytes * 8, out, (uint64_t *)Xi, ivec, key);
63 }
64 break;
65 case 12:
66 if (IS_CPU_SUPPORT_UNROLL8_EOR3()) {
67 unroll8_eor3_aes_gcm_dec_192_kernel(in, align_bytes * 8, out, (uint64_t *)Xi, ivec, key);
68 } else {
69 aes_gcm_dec_192_kernel(in, align_bytes * 8, out, (uint64_t *)Xi, ivec, key);
70 }
71 break;
72 case 14:
73 if (IS_CPU_SUPPORT_UNROLL8_EOR3()) {
74 unroll8_eor3_aes_gcm_dec_256_kernel(in, align_bytes * 8, out, (uint64_t *)Xi, ivec, key);
75 } else {
76 aes_gcm_dec_256_kernel(in, align_bytes * 8, out, (uint64_t *)Xi, ivec, key);
77 }
78 break;
79 }
80 return align_bytes;
81 }
82
83 static int armv8_aes_gcm_initkey(PROV_GCM_CTX *ctx, const unsigned char *key,
84 size_t keylen)
85 {
86 PROV_AES_GCM_CTX *actx = (PROV_AES_GCM_CTX *)ctx;
87 AES_KEY *ks = &actx->ks.ks;
88
89 GCM_HW_SET_KEY_CTR_FN(ks, aes_v8_set_encrypt_key, aes_v8_encrypt,
90 aes_v8_ctr32_encrypt_blocks);
91 return 1;
92 }
93
94
95 static const PROV_GCM_HW armv8_aes_gcm = {
96 armv8_aes_gcm_initkey,
97 ossl_gcm_setiv,
98 ossl_gcm_aad_update,
99 generic_aes_gcm_cipher_update,
100 ossl_gcm_cipher_final,
101 ossl_gcm_one_shot
102 };
103
104 const PROV_GCM_HW *ossl_prov_aes_hw_gcm(size_t keybits)
105 {
106 return AES_PMULL_CAPABLE ? &armv8_aes_gcm : &aes_gcm;
107 }