1 // SPDX-License-Identifier: GPL-2.0-or-later << 2 /* 1 /* 3 * Cryptographic API. 2 * Cryptographic API. 4 * 3 * 5 * Single-block cipher operations. !! 4 * Cipher operations. 6 * 5 * 7 * Copyright (c) 2002 James Morris <jmorris@in 6 * Copyright (c) 2002 James Morris <jmorris@intercode.com.au> 8 * Copyright (c) 2005 Herbert Xu <herbert@gond 7 * Copyright (c) 2005 Herbert Xu <herbert@gondor.apana.org.au> >> 8 * >> 9 * This program is free software; you can redistribute it and/or modify it >> 10 * under the terms of the GNU General Public License as published by the Free >> 11 * Software Foundation; either version 2 of the License, or (at your option) >> 12 * any later version. >> 13 * 9 */ 14 */ 10 15 11 #include <crypto/algapi.h> << 12 #include <crypto/internal/cipher.h> << 13 #include <linux/kernel.h> 16 #include <linux/kernel.h> 14 #include <linux/crypto.h> 17 #include <linux/crypto.h> 15 #include <linux/errno.h> 18 #include <linux/errno.h> 16 #include <linux/slab.h> 19 #include <linux/slab.h> 17 #include <linux/string.h> 20 #include <linux/string.h> 18 #include "internal.h" 21 #include "internal.h" 19 22 20 static int setkey_unaligned(struct crypto_ciph !! 23 static int setkey_unaligned(struct crypto_tfm *tfm, const u8 *key, 21 unsigned int keyle 24 unsigned int keylen) 22 { 25 { 23 struct cipher_alg *cia = crypto_cipher !! 26 struct cipher_alg *cia = &tfm->__crt_alg->cra_cipher; 24 unsigned long alignmask = crypto_ciphe !! 27 unsigned long alignmask = crypto_tfm_alg_alignmask(tfm); 25 int ret; 28 int ret; 26 u8 *buffer, *alignbuffer; 29 u8 *buffer, *alignbuffer; 27 unsigned long absize; 30 unsigned long absize; 28 31 29 absize = keylen + alignmask; 32 absize = keylen + alignmask; 30 buffer = kmalloc(absize, GFP_ATOMIC); 33 buffer = kmalloc(absize, GFP_ATOMIC); 31 if (!buffer) 34 if (!buffer) 32 return -ENOMEM; 35 return -ENOMEM; 33 36 34 alignbuffer = (u8 *)ALIGN((unsigned lo 37 alignbuffer = (u8 *)ALIGN((unsigned long)buffer, alignmask + 1); 35 memcpy(alignbuffer, key, keylen); 38 memcpy(alignbuffer, key, keylen); 36 ret = cia->cia_setkey(crypto_cipher_tf !! 39 ret = cia->cia_setkey(tfm, alignbuffer, keylen); 37 kfree_sensitive(buffer); !! 40 memset(alignbuffer, 0, keylen); >> 41 kfree(buffer); 38 return ret; 42 return ret; 39 43 40 } 44 } 41 45 42 int crypto_cipher_setkey(struct crypto_cipher !! 46 static int setkey(struct crypto_tfm *tfm, const u8 *key, unsigned int keylen) 43 const u8 *key, unsign << 44 { 47 { 45 struct cipher_alg *cia = crypto_cipher !! 48 struct cipher_alg *cia = &tfm->__crt_alg->cra_cipher; 46 unsigned long alignmask = crypto_ciphe !! 49 unsigned long alignmask = crypto_tfm_alg_alignmask(tfm); 47 50 48 if (keylen < cia->cia_min_keysize || k !! 51 tfm->crt_flags &= ~CRYPTO_TFM_RES_MASK; >> 52 if (keylen < cia->cia_min_keysize || keylen > cia->cia_max_keysize) { >> 53 tfm->crt_flags |= CRYPTO_TFM_RES_BAD_KEY_LEN; 49 return -EINVAL; 54 return -EINVAL; >> 55 } 50 56 51 if ((unsigned long)key & alignmask) 57 if ((unsigned long)key & alignmask) 52 return setkey_unaligned(tfm, k 58 return setkey_unaligned(tfm, key, keylen); 53 59 54 return cia->cia_setkey(crypto_cipher_t !! 60 return cia->cia_setkey(tfm, key, keylen); 55 } 61 } 56 EXPORT_SYMBOL_NS_GPL(crypto_cipher_setkey, CRY << 57 62 58 static inline void cipher_crypt_one(struct cry !! 63 static void cipher_crypt_unaligned(void (*fn)(struct crypto_tfm *, u8 *, 59 u8 *dst, c !! 64 const u8 *), >> 65 struct crypto_tfm *tfm, >> 66 u8 *dst, const u8 *src) 60 { 67 { 61 unsigned long alignmask = crypto_ciphe !! 68 unsigned long alignmask = crypto_tfm_alg_alignmask(tfm); 62 struct cipher_alg *cia = crypto_cipher !! 69 unsigned int size = crypto_tfm_alg_blocksize(tfm); 63 void (*fn)(struct crypto_tfm *, u8 *, !! 70 u8 buffer[size + alignmask]; 64 enc ? cia->cia_encrypt : cia-> !! 71 u8 *tmp = (u8 *)ALIGN((unsigned long)buffer, alignmask + 1); 65 72 66 if (unlikely(((unsigned long)dst | (un !! 73 memcpy(tmp, src, size); 67 unsigned int bs = crypto_ciphe !! 74 fn(tfm, tmp, tmp); 68 u8 buffer[MAX_CIPHER_BLOCKSIZE !! 75 memcpy(dst, tmp, size); 69 u8 *tmp = (u8 *)ALIGN((unsigne << 70 << 71 memcpy(tmp, src, bs); << 72 fn(crypto_cipher_tfm(tfm), tmp << 73 memcpy(dst, tmp, bs); << 74 } else { << 75 fn(crypto_cipher_tfm(tfm), dst << 76 } << 77 } 76 } 78 77 79 void crypto_cipher_encrypt_one(struct crypto_c !! 78 static void cipher_encrypt_unaligned(struct crypto_tfm *tfm, 80 u8 *dst, const !! 79 u8 *dst, const u8 *src) 81 { 80 { 82 cipher_crypt_one(tfm, dst, src, true); !! 81 unsigned long alignmask = crypto_tfm_alg_alignmask(tfm); 83 } !! 82 struct cipher_alg *cipher = &tfm->__crt_alg->cra_cipher; 84 EXPORT_SYMBOL_NS_GPL(crypto_cipher_encrypt_one << 85 83 86 void crypto_cipher_decrypt_one(struct crypto_c !! 84 if (unlikely(((unsigned long)dst | (unsigned long)src) & alignmask)) { 87 u8 *dst, const !! 85 cipher_crypt_unaligned(cipher->cia_encrypt, tfm, dst, src); 88 { !! 86 return; 89 cipher_crypt_one(tfm, dst, src, false) !! 87 } >> 88 >> 89 cipher->cia_encrypt(tfm, dst, src); 90 } 90 } 91 EXPORT_SYMBOL_NS_GPL(crypto_cipher_decrypt_one << 92 91 93 struct crypto_cipher *crypto_clone_cipher(stru !! 92 static void cipher_decrypt_unaligned(struct crypto_tfm *tfm, >> 93 u8 *dst, const u8 *src) 94 { 94 { 95 struct crypto_tfm *tfm = crypto_cipher !! 95 unsigned long alignmask = crypto_tfm_alg_alignmask(tfm); 96 struct crypto_alg *alg = tfm->__crt_al !! 96 struct cipher_alg *cipher = &tfm->__crt_alg->cra_cipher; 97 struct crypto_cipher *ncipher; << 98 struct crypto_tfm *ntfm; << 99 97 100 if (alg->cra_init) !! 98 if (unlikely(((unsigned long)dst | (unsigned long)src) & alignmask)) { 101 return ERR_PTR(-ENOSYS); !! 99 cipher_crypt_unaligned(cipher->cia_decrypt, tfm, dst, src); >> 100 return; >> 101 } 102 102 103 if (unlikely(!crypto_mod_get(alg))) !! 103 cipher->cia_decrypt(tfm, dst, src); 104 return ERR_PTR(-ESTALE); !! 104 } 105 105 106 ntfm = __crypto_alloc_tfmgfp(alg, CRYP !! 106 int crypto_init_cipher_ops(struct crypto_tfm *tfm) 107 CRYPTO_AL !! 107 { 108 if (IS_ERR(ntfm)) { !! 108 struct cipher_tfm *ops = &tfm->crt_cipher; 109 crypto_mod_put(alg); !! 109 struct cipher_alg *cipher = &tfm->__crt_alg->cra_cipher; 110 return ERR_CAST(ntfm); << 111 } << 112 110 113 ntfm->crt_flags = tfm->crt_flags; !! 111 ops->cit_setkey = setkey; >> 112 ops->cit_encrypt_one = crypto_tfm_alg_alignmask(tfm) ? >> 113 cipher_encrypt_unaligned : cipher->cia_encrypt; >> 114 ops->cit_decrypt_one = crypto_tfm_alg_alignmask(tfm) ? >> 115 cipher_decrypt_unaligned : cipher->cia_decrypt; 114 116 115 ncipher = __crypto_cipher_cast(ntfm); !! 117 return 0; >> 118 } 116 119 117 return ncipher; !! 120 void crypto_exit_cipher_ops(struct crypto_tfm *tfm) >> 121 { 118 } 122 } 119 EXPORT_SYMBOL_GPL(crypto_clone_cipher); << 120 123
Linux® is a registered trademark of Linus Torvalds in the United States and other countries.
TOMOYO® is a registered trademark of NTT DATA CORPORATION.