1 // SPDX-License-Identifier: GPL-2.0-or-later 1 // SPDX-License-Identifier: GPL-2.0-or-later 2 /* 2 /* 3 * Cryptographic API. 3 * Cryptographic API. 4 * 4 * 5 * Single-block cipher operations. 5 * Single-block cipher operations. 6 * 6 * 7 * Copyright (c) 2002 James Morris <jmorris@in 7 * Copyright (c) 2002 James Morris <jmorris@intercode.com.au> 8 * Copyright (c) 2005 Herbert Xu <herbert@gond 8 * Copyright (c) 2005 Herbert Xu <herbert@gondor.apana.org.au> 9 */ 9 */ 10 10 11 #include <crypto/algapi.h> 11 #include <crypto/algapi.h> 12 #include <crypto/internal/cipher.h> << 13 #include <linux/kernel.h> 12 #include <linux/kernel.h> 14 #include <linux/crypto.h> 13 #include <linux/crypto.h> 15 #include <linux/errno.h> 14 #include <linux/errno.h> 16 #include <linux/slab.h> 15 #include <linux/slab.h> 17 #include <linux/string.h> 16 #include <linux/string.h> 18 #include "internal.h" 17 #include "internal.h" 19 18 20 static int setkey_unaligned(struct crypto_ciph 19 static int setkey_unaligned(struct crypto_cipher *tfm, const u8 *key, 21 unsigned int keyle 20 unsigned int keylen) 22 { 21 { 23 struct cipher_alg *cia = crypto_cipher 22 struct cipher_alg *cia = crypto_cipher_alg(tfm); 24 unsigned long alignmask = crypto_ciphe 23 unsigned long alignmask = crypto_cipher_alignmask(tfm); 25 int ret; 24 int ret; 26 u8 *buffer, *alignbuffer; 25 u8 *buffer, *alignbuffer; 27 unsigned long absize; 26 unsigned long absize; 28 27 29 absize = keylen + alignmask; 28 absize = keylen + alignmask; 30 buffer = kmalloc(absize, GFP_ATOMIC); 29 buffer = kmalloc(absize, GFP_ATOMIC); 31 if (!buffer) 30 if (!buffer) 32 return -ENOMEM; 31 return -ENOMEM; 33 32 34 alignbuffer = (u8 *)ALIGN((unsigned lo 33 alignbuffer = (u8 *)ALIGN((unsigned long)buffer, alignmask + 1); 35 memcpy(alignbuffer, key, keylen); 34 memcpy(alignbuffer, key, keylen); 36 ret = cia->cia_setkey(crypto_cipher_tf 35 ret = cia->cia_setkey(crypto_cipher_tfm(tfm), alignbuffer, keylen); 37 kfree_sensitive(buffer); !! 36 memset(alignbuffer, 0, keylen); >> 37 kfree(buffer); 38 return ret; 38 return ret; 39 39 40 } 40 } 41 41 42 int crypto_cipher_setkey(struct crypto_cipher 42 int crypto_cipher_setkey(struct crypto_cipher *tfm, 43 const u8 *key, unsign 43 const u8 *key, unsigned int keylen) 44 { 44 { 45 struct cipher_alg *cia = crypto_cipher 45 struct cipher_alg *cia = crypto_cipher_alg(tfm); 46 unsigned long alignmask = crypto_ciphe 46 unsigned long alignmask = crypto_cipher_alignmask(tfm); 47 47 48 if (keylen < cia->cia_min_keysize || k 48 if (keylen < cia->cia_min_keysize || keylen > cia->cia_max_keysize) 49 return -EINVAL; 49 return -EINVAL; 50 50 51 if ((unsigned long)key & alignmask) 51 if ((unsigned long)key & alignmask) 52 return setkey_unaligned(tfm, k 52 return setkey_unaligned(tfm, key, keylen); 53 53 54 return cia->cia_setkey(crypto_cipher_t 54 return cia->cia_setkey(crypto_cipher_tfm(tfm), key, keylen); 55 } 55 } 56 EXPORT_SYMBOL_NS_GPL(crypto_cipher_setkey, CRY !! 56 EXPORT_SYMBOL_GPL(crypto_cipher_setkey); 57 57 58 static inline void cipher_crypt_one(struct cry 58 static inline void cipher_crypt_one(struct crypto_cipher *tfm, 59 u8 *dst, c 59 u8 *dst, const u8 *src, bool enc) 60 { 60 { 61 unsigned long alignmask = crypto_ciphe 61 unsigned long alignmask = crypto_cipher_alignmask(tfm); 62 struct cipher_alg *cia = crypto_cipher 62 struct cipher_alg *cia = crypto_cipher_alg(tfm); 63 void (*fn)(struct crypto_tfm *, u8 *, 63 void (*fn)(struct crypto_tfm *, u8 *, const u8 *) = 64 enc ? cia->cia_encrypt : cia-> 64 enc ? cia->cia_encrypt : cia->cia_decrypt; 65 65 66 if (unlikely(((unsigned long)dst | (un 66 if (unlikely(((unsigned long)dst | (unsigned long)src) & alignmask)) { 67 unsigned int bs = crypto_ciphe 67 unsigned int bs = crypto_cipher_blocksize(tfm); 68 u8 buffer[MAX_CIPHER_BLOCKSIZE 68 u8 buffer[MAX_CIPHER_BLOCKSIZE + MAX_CIPHER_ALIGNMASK]; 69 u8 *tmp = (u8 *)ALIGN((unsigne 69 u8 *tmp = (u8 *)ALIGN((unsigned long)buffer, alignmask + 1); 70 70 71 memcpy(tmp, src, bs); 71 memcpy(tmp, src, bs); 72 fn(crypto_cipher_tfm(tfm), tmp 72 fn(crypto_cipher_tfm(tfm), tmp, tmp); 73 memcpy(dst, tmp, bs); 73 memcpy(dst, tmp, bs); 74 } else { 74 } else { 75 fn(crypto_cipher_tfm(tfm), dst 75 fn(crypto_cipher_tfm(tfm), dst, src); 76 } 76 } 77 } 77 } 78 78 79 void crypto_cipher_encrypt_one(struct crypto_c 79 void crypto_cipher_encrypt_one(struct crypto_cipher *tfm, 80 u8 *dst, const 80 u8 *dst, const u8 *src) 81 { 81 { 82 cipher_crypt_one(tfm, dst, src, true); 82 cipher_crypt_one(tfm, dst, src, true); 83 } 83 } 84 EXPORT_SYMBOL_NS_GPL(crypto_cipher_encrypt_one !! 84 EXPORT_SYMBOL_GPL(crypto_cipher_encrypt_one); 85 85 86 void crypto_cipher_decrypt_one(struct crypto_c 86 void crypto_cipher_decrypt_one(struct crypto_cipher *tfm, 87 u8 *dst, const 87 u8 *dst, const u8 *src) 88 { 88 { 89 cipher_crypt_one(tfm, dst, src, false) 89 cipher_crypt_one(tfm, dst, src, false); 90 } 90 } 91 EXPORT_SYMBOL_NS_GPL(crypto_cipher_decrypt_one !! 91 EXPORT_SYMBOL_GPL(crypto_cipher_decrypt_one); 92 << 93 struct crypto_cipher *crypto_clone_cipher(stru << 94 { << 95 struct crypto_tfm *tfm = crypto_cipher << 96 struct crypto_alg *alg = tfm->__crt_al << 97 struct crypto_cipher *ncipher; << 98 struct crypto_tfm *ntfm; << 99 << 100 if (alg->cra_init) << 101 return ERR_PTR(-ENOSYS); << 102 << 103 if (unlikely(!crypto_mod_get(alg))) << 104 return ERR_PTR(-ESTALE); << 105 << 106 ntfm = __crypto_alloc_tfmgfp(alg, CRYP << 107 CRYPTO_AL << 108 if (IS_ERR(ntfm)) { << 109 crypto_mod_put(alg); << 110 return ERR_CAST(ntfm); << 111 } << 112 << 113 ntfm->crt_flags = tfm->crt_flags; << 114 << 115 ncipher = __crypto_cipher_cast(ntfm); << 116 << 117 return ncipher; << 118 } << 119 EXPORT_SYMBOL_GPL(crypto_clone_cipher); << 120 92
Linux® is a registered trademark of Linus Torvalds in the United States and other countries.
TOMOYO® is a registered trademark of NTT DATA CORPORATION.