~ [ source navigation ] ~ [ diff markup ] ~ [ identifier search ] ~

TOMOYO Linux Cross Reference
Linux/arch/sparc/crypto/aes_glue.c

Version: ~ [ linux-6.11-rc3 ] ~ [ linux-6.10.4 ] ~ [ linux-6.9.12 ] ~ [ linux-6.8.12 ] ~ [ linux-6.7.12 ] ~ [ linux-6.6.45 ] ~ [ linux-6.5.13 ] ~ [ linux-6.4.16 ] ~ [ linux-6.3.13 ] ~ [ linux-6.2.16 ] ~ [ linux-6.1.104 ] ~ [ linux-6.0.19 ] ~ [ linux-5.19.17 ] ~ [ linux-5.18.19 ] ~ [ linux-5.17.15 ] ~ [ linux-5.16.20 ] ~ [ linux-5.15.164 ] ~ [ linux-5.14.21 ] ~ [ linux-5.13.19 ] ~ [ linux-5.12.19 ] ~ [ linux-5.11.22 ] ~ [ linux-5.10.223 ] ~ [ linux-5.9.16 ] ~ [ linux-5.8.18 ] ~ [ linux-5.7.19 ] ~ [ linux-5.6.19 ] ~ [ linux-5.5.19 ] ~ [ linux-5.4.281 ] ~ [ linux-5.3.18 ] ~ [ linux-5.2.21 ] ~ [ linux-5.1.21 ] ~ [ linux-5.0.21 ] ~ [ linux-4.20.17 ] ~ [ linux-4.19.319 ] ~ [ linux-4.18.20 ] ~ [ linux-4.17.19 ] ~ [ linux-4.16.18 ] ~ [ linux-4.15.18 ] ~ [ linux-4.14.336 ] ~ [ linux-4.13.16 ] ~ [ linux-4.12.14 ] ~ [ linux-4.11.12 ] ~ [ linux-4.10.17 ] ~ [ linux-4.9.337 ] ~ [ linux-4.4.302 ] ~ [ linux-3.10.108 ] ~ [ linux-2.6.32.71 ] ~ [ linux-2.6.0 ] ~ [ linux-2.4.37.11 ] ~ [ unix-v6-master ] ~ [ ccs-tools-1.8.9 ] ~ [ policy-sample ] ~
Architecture: ~ [ i386 ] ~ [ alpha ] ~ [ m68k ] ~ [ mips ] ~ [ ppc ] ~ [ sparc ] ~ [ sparc64 ] ~

  1 // SPDX-License-Identifier: GPL-2.0-only
  2 /* Glue code for AES encryption optimized for sparc64 crypto opcodes.
  3  *
  4  * This is based largely upon arch/x86/crypto/aesni-intel_glue.c
  5  *
  6  * Copyright (C) 2008, Intel Corp.
  7  *    Author: Huang Ying <ying.huang@intel.com>
  8  *
  9  * Added RFC4106 AES-GCM support for 128-bit keys under the AEAD
 10  * interface for 64-bit kernels.
 11  *    Authors: Adrian Hoban <adrian.hoban@intel.com>
 12  *             Gabriele Paoloni <gabriele.paoloni@intel.com>
 13  *             Tadeusz Struk (tadeusz.struk@intel.com)
 14  *             Aidan O'Mahony (aidan.o.mahony@intel.com)
 15  *    Copyright (c) 2010, Intel Corporation.
 16  */
 17 
 18 #define pr_fmt(fmt)     KBUILD_MODNAME ": " fmt
 19 
 20 #include <linux/crypto.h>
 21 #include <linux/init.h>
 22 #include <linux/module.h>
 23 #include <linux/mm.h>
 24 #include <linux/types.h>
 25 #include <crypto/algapi.h>
 26 #include <crypto/aes.h>
 27 #include <crypto/internal/skcipher.h>
 28 
 29 #include <asm/fpumacro.h>
 30 #include <asm/pstate.h>
 31 #include <asm/elf.h>
 32 
 33 #include "opcodes.h"
 34 
 35 struct aes_ops {
 36         void (*encrypt)(const u64 *key, const u32 *input, u32 *output);
 37         void (*decrypt)(const u64 *key, const u32 *input, u32 *output);
 38         void (*load_encrypt_keys)(const u64 *key);
 39         void (*load_decrypt_keys)(const u64 *key);
 40         void (*ecb_encrypt)(const u64 *key, const u64 *input, u64 *output,
 41                             unsigned int len);
 42         void (*ecb_decrypt)(const u64 *key, const u64 *input, u64 *output,
 43                             unsigned int len);
 44         void (*cbc_encrypt)(const u64 *key, const u64 *input, u64 *output,
 45                             unsigned int len, u64 *iv);
 46         void (*cbc_decrypt)(const u64 *key, const u64 *input, u64 *output,
 47                             unsigned int len, u64 *iv);
 48         void (*ctr_crypt)(const u64 *key, const u64 *input, u64 *output,
 49                           unsigned int len, u64 *iv);
 50 };
 51 
 52 struct crypto_sparc64_aes_ctx {
 53         struct aes_ops *ops;
 54         u64 key[AES_MAX_KEYLENGTH / sizeof(u64)];
 55         u32 key_length;
 56         u32 expanded_key_length;
 57 };
 58 
 59 extern void aes_sparc64_encrypt_128(const u64 *key, const u32 *input,
 60                                     u32 *output);
 61 extern void aes_sparc64_encrypt_192(const u64 *key, const u32 *input,
 62                                     u32 *output);
 63 extern void aes_sparc64_encrypt_256(const u64 *key, const u32 *input,
 64                                     u32 *output);
 65 
 66 extern void aes_sparc64_decrypt_128(const u64 *key, const u32 *input,
 67                                     u32 *output);
 68 extern void aes_sparc64_decrypt_192(const u64 *key, const u32 *input,
 69                                     u32 *output);
 70 extern void aes_sparc64_decrypt_256(const u64 *key, const u32 *input,
 71                                     u32 *output);
 72 
 73 extern void aes_sparc64_load_encrypt_keys_128(const u64 *key);
 74 extern void aes_sparc64_load_encrypt_keys_192(const u64 *key);
 75 extern void aes_sparc64_load_encrypt_keys_256(const u64 *key);
 76 
 77 extern void aes_sparc64_load_decrypt_keys_128(const u64 *key);
 78 extern void aes_sparc64_load_decrypt_keys_192(const u64 *key);
 79 extern void aes_sparc64_load_decrypt_keys_256(const u64 *key);
 80 
 81 extern void aes_sparc64_ecb_encrypt_128(const u64 *key, const u64 *input,
 82                                         u64 *output, unsigned int len);
 83 extern void aes_sparc64_ecb_encrypt_192(const u64 *key, const u64 *input,
 84                                         u64 *output, unsigned int len);
 85 extern void aes_sparc64_ecb_encrypt_256(const u64 *key, const u64 *input,
 86                                         u64 *output, unsigned int len);
 87 
 88 extern void aes_sparc64_ecb_decrypt_128(const u64 *key, const u64 *input,
 89                                         u64 *output, unsigned int len);
 90 extern void aes_sparc64_ecb_decrypt_192(const u64 *key, const u64 *input,
 91                                         u64 *output, unsigned int len);
 92 extern void aes_sparc64_ecb_decrypt_256(const u64 *key, const u64 *input,
 93                                         u64 *output, unsigned int len);
 94 
 95 extern void aes_sparc64_cbc_encrypt_128(const u64 *key, const u64 *input,
 96                                         u64 *output, unsigned int len,
 97                                         u64 *iv);
 98 
 99 extern void aes_sparc64_cbc_encrypt_192(const u64 *key, const u64 *input,
100                                         u64 *output, unsigned int len,
101                                         u64 *iv);
102 
103 extern void aes_sparc64_cbc_encrypt_256(const u64 *key, const u64 *input,
104                                         u64 *output, unsigned int len,
105                                         u64 *iv);
106 
107 extern void aes_sparc64_cbc_decrypt_128(const u64 *key, const u64 *input,
108                                         u64 *output, unsigned int len,
109                                         u64 *iv);
110 
111 extern void aes_sparc64_cbc_decrypt_192(const u64 *key, const u64 *input,
112                                         u64 *output, unsigned int len,
113                                         u64 *iv);
114 
115 extern void aes_sparc64_cbc_decrypt_256(const u64 *key, const u64 *input,
116                                         u64 *output, unsigned int len,
117                                         u64 *iv);
118 
119 extern void aes_sparc64_ctr_crypt_128(const u64 *key, const u64 *input,
120                                       u64 *output, unsigned int len,
121                                       u64 *iv);
122 extern void aes_sparc64_ctr_crypt_192(const u64 *key, const u64 *input,
123                                       u64 *output, unsigned int len,
124                                       u64 *iv);
125 extern void aes_sparc64_ctr_crypt_256(const u64 *key, const u64 *input,
126                                       u64 *output, unsigned int len,
127                                       u64 *iv);
128 
129 static struct aes_ops aes128_ops = {
130         .encrypt                = aes_sparc64_encrypt_128,
131         .decrypt                = aes_sparc64_decrypt_128,
132         .load_encrypt_keys      = aes_sparc64_load_encrypt_keys_128,
133         .load_decrypt_keys      = aes_sparc64_load_decrypt_keys_128,
134         .ecb_encrypt            = aes_sparc64_ecb_encrypt_128,
135         .ecb_decrypt            = aes_sparc64_ecb_decrypt_128,
136         .cbc_encrypt            = aes_sparc64_cbc_encrypt_128,
137         .cbc_decrypt            = aes_sparc64_cbc_decrypt_128,
138         .ctr_crypt              = aes_sparc64_ctr_crypt_128,
139 };
140 
141 static struct aes_ops aes192_ops = {
142         .encrypt                = aes_sparc64_encrypt_192,
143         .decrypt                = aes_sparc64_decrypt_192,
144         .load_encrypt_keys      = aes_sparc64_load_encrypt_keys_192,
145         .load_decrypt_keys      = aes_sparc64_load_decrypt_keys_192,
146         .ecb_encrypt            = aes_sparc64_ecb_encrypt_192,
147         .ecb_decrypt            = aes_sparc64_ecb_decrypt_192,
148         .cbc_encrypt            = aes_sparc64_cbc_encrypt_192,
149         .cbc_decrypt            = aes_sparc64_cbc_decrypt_192,
150         .ctr_crypt              = aes_sparc64_ctr_crypt_192,
151 };
152 
153 static struct aes_ops aes256_ops = {
154         .encrypt                = aes_sparc64_encrypt_256,
155         .decrypt                = aes_sparc64_decrypt_256,
156         .load_encrypt_keys      = aes_sparc64_load_encrypt_keys_256,
157         .load_decrypt_keys      = aes_sparc64_load_decrypt_keys_256,
158         .ecb_encrypt            = aes_sparc64_ecb_encrypt_256,
159         .ecb_decrypt            = aes_sparc64_ecb_decrypt_256,
160         .cbc_encrypt            = aes_sparc64_cbc_encrypt_256,
161         .cbc_decrypt            = aes_sparc64_cbc_decrypt_256,
162         .ctr_crypt              = aes_sparc64_ctr_crypt_256,
163 };
164 
165 extern void aes_sparc64_key_expand(const u32 *in_key, u64 *output_key,
166                                    unsigned int key_len);
167 
168 static int aes_set_key(struct crypto_tfm *tfm, const u8 *in_key,
169                        unsigned int key_len)
170 {
171         struct crypto_sparc64_aes_ctx *ctx = crypto_tfm_ctx(tfm);
172 
173         switch (key_len) {
174         case AES_KEYSIZE_128:
175                 ctx->expanded_key_length = 0xb0;
176                 ctx->ops = &aes128_ops;
177                 break;
178 
179         case AES_KEYSIZE_192:
180                 ctx->expanded_key_length = 0xd0;
181                 ctx->ops = &aes192_ops;
182                 break;
183 
184         case AES_KEYSIZE_256:
185                 ctx->expanded_key_length = 0xf0;
186                 ctx->ops = &aes256_ops;
187                 break;
188 
189         default:
190                 return -EINVAL;
191         }
192 
193         aes_sparc64_key_expand((const u32 *)in_key, &ctx->key[0], key_len);
194         ctx->key_length = key_len;
195 
196         return 0;
197 }
198 
199 static int aes_set_key_skcipher(struct crypto_skcipher *tfm, const u8 *in_key,
200                                 unsigned int key_len)
201 {
202         return aes_set_key(crypto_skcipher_tfm(tfm), in_key, key_len);
203 }
204 
205 static void crypto_aes_encrypt(struct crypto_tfm *tfm, u8 *dst, const u8 *src)
206 {
207         struct crypto_sparc64_aes_ctx *ctx = crypto_tfm_ctx(tfm);
208 
209         ctx->ops->encrypt(&ctx->key[0], (const u32 *) src, (u32 *) dst);
210 }
211 
212 static void crypto_aes_decrypt(struct crypto_tfm *tfm, u8 *dst, const u8 *src)
213 {
214         struct crypto_sparc64_aes_ctx *ctx = crypto_tfm_ctx(tfm);
215 
216         ctx->ops->decrypt(&ctx->key[0], (const u32 *) src, (u32 *) dst);
217 }
218 
219 static int ecb_encrypt(struct skcipher_request *req)
220 {
221         struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(req);
222         const struct crypto_sparc64_aes_ctx *ctx = crypto_skcipher_ctx(tfm);
223         struct skcipher_walk walk;
224         unsigned int nbytes;
225         int err;
226 
227         err = skcipher_walk_virt(&walk, req, true);
228         if (err)
229                 return err;
230 
231         ctx->ops->load_encrypt_keys(&ctx->key[0]);
232         while ((nbytes = walk.nbytes) != 0) {
233                 ctx->ops->ecb_encrypt(&ctx->key[0], walk.src.virt.addr,
234                                       walk.dst.virt.addr,
235                                       round_down(nbytes, AES_BLOCK_SIZE));
236                 err = skcipher_walk_done(&walk, nbytes % AES_BLOCK_SIZE);
237         }
238         fprs_write(0);
239         return err;
240 }
241 
242 static int ecb_decrypt(struct skcipher_request *req)
243 {
244         struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(req);
245         const struct crypto_sparc64_aes_ctx *ctx = crypto_skcipher_ctx(tfm);
246         const u64 *key_end;
247         struct skcipher_walk walk;
248         unsigned int nbytes;
249         int err;
250 
251         err = skcipher_walk_virt(&walk, req, true);
252         if (err)
253                 return err;
254 
255         ctx->ops->load_decrypt_keys(&ctx->key[0]);
256         key_end = &ctx->key[ctx->expanded_key_length / sizeof(u64)];
257         while ((nbytes = walk.nbytes) != 0) {
258                 ctx->ops->ecb_decrypt(key_end, walk.src.virt.addr,
259                                       walk.dst.virt.addr,
260                                       round_down(nbytes, AES_BLOCK_SIZE));
261                 err = skcipher_walk_done(&walk, nbytes % AES_BLOCK_SIZE);
262         }
263         fprs_write(0);
264 
265         return err;
266 }
267 
268 static int cbc_encrypt(struct skcipher_request *req)
269 {
270         struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(req);
271         const struct crypto_sparc64_aes_ctx *ctx = crypto_skcipher_ctx(tfm);
272         struct skcipher_walk walk;
273         unsigned int nbytes;
274         int err;
275 
276         err = skcipher_walk_virt(&walk, req, true);
277         if (err)
278                 return err;
279 
280         ctx->ops->load_encrypt_keys(&ctx->key[0]);
281         while ((nbytes = walk.nbytes) != 0) {
282                 ctx->ops->cbc_encrypt(&ctx->key[0], walk.src.virt.addr,
283                                       walk.dst.virt.addr,
284                                       round_down(nbytes, AES_BLOCK_SIZE),
285                                       walk.iv);
286                 err = skcipher_walk_done(&walk, nbytes % AES_BLOCK_SIZE);
287         }
288         fprs_write(0);
289         return err;
290 }
291 
292 static int cbc_decrypt(struct skcipher_request *req)
293 {
294         struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(req);
295         const struct crypto_sparc64_aes_ctx *ctx = crypto_skcipher_ctx(tfm);
296         const u64 *key_end;
297         struct skcipher_walk walk;
298         unsigned int nbytes;
299         int err;
300 
301         err = skcipher_walk_virt(&walk, req, true);
302         if (err)
303                 return err;
304 
305         ctx->ops->load_decrypt_keys(&ctx->key[0]);
306         key_end = &ctx->key[ctx->expanded_key_length / sizeof(u64)];
307         while ((nbytes = walk.nbytes) != 0) {
308                 ctx->ops->cbc_decrypt(key_end, walk.src.virt.addr,
309                                       walk.dst.virt.addr,
310                                       round_down(nbytes, AES_BLOCK_SIZE),
311                                       walk.iv);
312                 err = skcipher_walk_done(&walk, nbytes % AES_BLOCK_SIZE);
313         }
314         fprs_write(0);
315 
316         return err;
317 }
318 
319 static void ctr_crypt_final(const struct crypto_sparc64_aes_ctx *ctx,
320                             struct skcipher_walk *walk)
321 {
322         u8 *ctrblk = walk->iv;
323         u64 keystream[AES_BLOCK_SIZE / sizeof(u64)];
324         u8 *src = walk->src.virt.addr;
325         u8 *dst = walk->dst.virt.addr;
326         unsigned int nbytes = walk->nbytes;
327 
328         ctx->ops->ecb_encrypt(&ctx->key[0], (const u64 *)ctrblk,
329                               keystream, AES_BLOCK_SIZE);
330         crypto_xor_cpy(dst, (u8 *) keystream, src, nbytes);
331         crypto_inc(ctrblk, AES_BLOCK_SIZE);
332 }
333 
334 static int ctr_crypt(struct skcipher_request *req)
335 {
336         struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(req);
337         const struct crypto_sparc64_aes_ctx *ctx = crypto_skcipher_ctx(tfm);
338         struct skcipher_walk walk;
339         unsigned int nbytes;
340         int err;
341 
342         err = skcipher_walk_virt(&walk, req, true);
343         if (err)
344                 return err;
345 
346         ctx->ops->load_encrypt_keys(&ctx->key[0]);
347         while ((nbytes = walk.nbytes) >= AES_BLOCK_SIZE) {
348                 ctx->ops->ctr_crypt(&ctx->key[0], walk.src.virt.addr,
349                                     walk.dst.virt.addr,
350                                     round_down(nbytes, AES_BLOCK_SIZE),
351                                     walk.iv);
352                 err = skcipher_walk_done(&walk, nbytes % AES_BLOCK_SIZE);
353         }
354         if (walk.nbytes) {
355                 ctr_crypt_final(ctx, &walk);
356                 err = skcipher_walk_done(&walk, 0);
357         }
358         fprs_write(0);
359         return err;
360 }
361 
362 static struct crypto_alg cipher_alg = {
363         .cra_name               = "aes",
364         .cra_driver_name        = "aes-sparc64",
365         .cra_priority           = SPARC_CR_OPCODE_PRIORITY,
366         .cra_flags              = CRYPTO_ALG_TYPE_CIPHER,
367         .cra_blocksize          = AES_BLOCK_SIZE,
368         .cra_ctxsize            = sizeof(struct crypto_sparc64_aes_ctx),
369         .cra_alignmask          = 3,
370         .cra_module             = THIS_MODULE,
371         .cra_u  = {
372                 .cipher = {
373                         .cia_min_keysize        = AES_MIN_KEY_SIZE,
374                         .cia_max_keysize        = AES_MAX_KEY_SIZE,
375                         .cia_setkey             = aes_set_key,
376                         .cia_encrypt            = crypto_aes_encrypt,
377                         .cia_decrypt            = crypto_aes_decrypt
378                 }
379         }
380 };
381 
382 static struct skcipher_alg skcipher_algs[] = {
383         {
384                 .base.cra_name          = "ecb(aes)",
385                 .base.cra_driver_name   = "ecb-aes-sparc64",
386                 .base.cra_priority      = SPARC_CR_OPCODE_PRIORITY,
387                 .base.cra_blocksize     = AES_BLOCK_SIZE,
388                 .base.cra_ctxsize       = sizeof(struct crypto_sparc64_aes_ctx),
389                 .base.cra_alignmask     = 7,
390                 .base.cra_module        = THIS_MODULE,
391                 .min_keysize            = AES_MIN_KEY_SIZE,
392                 .max_keysize            = AES_MAX_KEY_SIZE,
393                 .setkey                 = aes_set_key_skcipher,
394                 .encrypt                = ecb_encrypt,
395                 .decrypt                = ecb_decrypt,
396         }, {
397                 .base.cra_name          = "cbc(aes)",
398                 .base.cra_driver_name   = "cbc-aes-sparc64",
399                 .base.cra_priority      = SPARC_CR_OPCODE_PRIORITY,
400                 .base.cra_blocksize     = AES_BLOCK_SIZE,
401                 .base.cra_ctxsize       = sizeof(struct crypto_sparc64_aes_ctx),
402                 .base.cra_alignmask     = 7,
403                 .base.cra_module        = THIS_MODULE,
404                 .min_keysize            = AES_MIN_KEY_SIZE,
405                 .max_keysize            = AES_MAX_KEY_SIZE,
406                 .ivsize                 = AES_BLOCK_SIZE,
407                 .setkey                 = aes_set_key_skcipher,
408                 .encrypt                = cbc_encrypt,
409                 .decrypt                = cbc_decrypt,
410         }, {
411                 .base.cra_name          = "ctr(aes)",
412                 .base.cra_driver_name   = "ctr-aes-sparc64",
413                 .base.cra_priority      = SPARC_CR_OPCODE_PRIORITY,
414                 .base.cra_blocksize     = 1,
415                 .base.cra_ctxsize       = sizeof(struct crypto_sparc64_aes_ctx),
416                 .base.cra_alignmask     = 7,
417                 .base.cra_module        = THIS_MODULE,
418                 .min_keysize            = AES_MIN_KEY_SIZE,
419                 .max_keysize            = AES_MAX_KEY_SIZE,
420                 .ivsize                 = AES_BLOCK_SIZE,
421                 .setkey                 = aes_set_key_skcipher,
422                 .encrypt                = ctr_crypt,
423                 .decrypt                = ctr_crypt,
424                 .chunksize              = AES_BLOCK_SIZE,
425         }
426 };
427 
428 static bool __init sparc64_has_aes_opcode(void)
429 {
430         unsigned long cfr;
431 
432         if (!(sparc64_elf_hwcap & HWCAP_SPARC_CRYPTO))
433                 return false;
434 
435         __asm__ __volatile__("rd %%asr26, %0" : "=r" (cfr));
436         if (!(cfr & CFR_AES))
437                 return false;
438 
439         return true;
440 }
441 
442 static int __init aes_sparc64_mod_init(void)
443 {
444         int err;
445 
446         if (!sparc64_has_aes_opcode()) {
447                 pr_info("sparc64 aes opcodes not available.\n");
448                 return -ENODEV;
449         }
450         pr_info("Using sparc64 aes opcodes optimized AES implementation\n");
451         err = crypto_register_alg(&cipher_alg);
452         if (err)
453                 return err;
454         err = crypto_register_skciphers(skcipher_algs,
455                                         ARRAY_SIZE(skcipher_algs));
456         if (err)
457                 crypto_unregister_alg(&cipher_alg);
458         return err;
459 }
460 
461 static void __exit aes_sparc64_mod_fini(void)
462 {
463         crypto_unregister_alg(&cipher_alg);
464         crypto_unregister_skciphers(skcipher_algs, ARRAY_SIZE(skcipher_algs));
465 }
466 
467 module_init(aes_sparc64_mod_init);
468 module_exit(aes_sparc64_mod_fini);
469 
470 MODULE_LICENSE("GPL");
471 MODULE_DESCRIPTION("Rijndael (AES) Cipher Algorithm, sparc64 aes opcode accelerated");
472 
473 MODULE_ALIAS_CRYPTO("aes");
474 
475 #include "crop_devid.c"
476 

~ [ source navigation ] ~ [ diff markup ] ~ [ identifier search ] ~

kernel.org | git.kernel.org | LWN.net | Project Home | SVN repository | Mail admin

Linux® is a registered trademark of Linus Torvalds in the United States and other countries.
TOMOYO® is a registered trademark of NTT DATA CORPORATION.

sflogo.php