]> git.ipfire.org Git - thirdparty/linux.git/commitdiff
crypto: khazad - stop using cra_alignmask
authorEric Biggers <ebiggers@google.com>
Sat, 7 Dec 2024 19:57:48 +0000 (11:57 -0800)
committerHerbert Xu <herbert@gondor.apana.org.au>
Sat, 14 Dec 2024 09:21:44 +0000 (17:21 +0800)
Instead of specifying a nonzero alignmask, use the unaligned access
helpers.  This eliminates unnecessary alignment operations on most CPUs,
which can handle unaligned accesses efficiently, and brings us a step
closer to eventually removing support for the alignmask field.

Signed-off-by: Eric Biggers <ebiggers@google.com>
Signed-off-by: Herbert Xu <herbert@gondor.apana.org.au>
crypto/khazad.c

index 70cafe73f97405ca52044f18446c2736fb4c7a8d..7ad338ca2c18fca6408e605af9a435fa49fc84b4 100644 (file)
@@ -23,7 +23,7 @@
 #include <linux/init.h>
 #include <linux/module.h>
 #include <linux/mm.h>
-#include <asm/byteorder.h>
+#include <linux/unaligned.h>
 #include <linux/types.h>
 
 #define KHAZAD_KEY_SIZE                16
@@ -757,14 +757,12 @@ static int khazad_setkey(struct crypto_tfm *tfm, const u8 *in_key,
                         unsigned int key_len)
 {
        struct khazad_ctx *ctx = crypto_tfm_ctx(tfm);
-       const __be32 *key = (const __be32 *)in_key;
        int r;
        const u64 *S = T7;
        u64 K2, K1;
 
-       /* key is supposed to be 32-bit aligned */
-       K2 = ((u64)be32_to_cpu(key[0]) << 32) | be32_to_cpu(key[1]);
-       K1 = ((u64)be32_to_cpu(key[2]) << 32) | be32_to_cpu(key[3]);
+       K2 = get_unaligned_be64(&in_key[0]);
+       K1 = get_unaligned_be64(&in_key[8]);
 
        /* setup the encrypt key */
        for (r = 0; r <= KHAZAD_ROUNDS; r++) {
@@ -800,14 +798,12 @@ static int khazad_setkey(struct crypto_tfm *tfm, const u8 *in_key,
 }
 
 static void khazad_crypt(const u64 roundKey[KHAZAD_ROUNDS + 1],
-               u8 *ciphertext, const u8 *plaintext)
+                        u8 *dst, const u8 *src)
 {
-       const __be64 *src = (const __be64 *)plaintext;
-       __be64 *dst = (__be64 *)ciphertext;
        int r;
        u64 state;
 
-       state = be64_to_cpu(*src) ^ roundKey[0];
+       state = get_unaligned_be64(src) ^ roundKey[0];
 
        for (r = 1; r < KHAZAD_ROUNDS; r++) {
                state = T0[(int)(state >> 56)       ] ^
@@ -831,7 +827,7 @@ static void khazad_crypt(const u64 roundKey[KHAZAD_ROUNDS + 1],
                (T7[(int)(state      ) & 0xff] & 0x00000000000000ffULL) ^
                roundKey[KHAZAD_ROUNDS];
 
-       *dst = cpu_to_be64(state);
+       put_unaligned_be64(state, dst);
 }
 
 static void khazad_encrypt(struct crypto_tfm *tfm, u8 *dst, const u8 *src)
@@ -852,7 +848,6 @@ static struct crypto_alg khazad_alg = {
        .cra_flags              =       CRYPTO_ALG_TYPE_CIPHER,
        .cra_blocksize          =       KHAZAD_BLOCK_SIZE,
        .cra_ctxsize            =       sizeof (struct khazad_ctx),
-       .cra_alignmask          =       7,
        .cra_module             =       THIS_MODULE,
        .cra_u                  =       { .cipher = {
        .cia_min_keysize        =       KHAZAD_KEY_SIZE,