~ [ source navigation ] ~ [ diff markup ] ~ [ identifier search ] ~

TOMOYO Linux Cross Reference
Linux/arch/x86/crypto/aes-gcm-avx10-x86_64.S

Version: ~ [ linux-6.12-rc7 ] ~ [ linux-6.11.7 ] ~ [ linux-6.10.14 ] ~ [ linux-6.9.12 ] ~ [ linux-6.8.12 ] ~ [ linux-6.7.12 ] ~ [ linux-6.6.60 ] ~ [ linux-6.5.13 ] ~ [ linux-6.4.16 ] ~ [ linux-6.3.13 ] ~ [ linux-6.2.16 ] ~ [ linux-6.1.116 ] ~ [ linux-6.0.19 ] ~ [ linux-5.19.17 ] ~ [ linux-5.18.19 ] ~ [ linux-5.17.15 ] ~ [ linux-5.16.20 ] ~ [ linux-5.15.171 ] ~ [ linux-5.14.21 ] ~ [ linux-5.13.19 ] ~ [ linux-5.12.19 ] ~ [ linux-5.11.22 ] ~ [ linux-5.10.229 ] ~ [ linux-5.9.16 ] ~ [ linux-5.8.18 ] ~ [ linux-5.7.19 ] ~ [ linux-5.6.19 ] ~ [ linux-5.5.19 ] ~ [ linux-5.4.285 ] ~ [ linux-5.3.18 ] ~ [ linux-5.2.21 ] ~ [ linux-5.1.21 ] ~ [ linux-5.0.21 ] ~ [ linux-4.20.17 ] ~ [ linux-4.19.323 ] ~ [ linux-4.18.20 ] ~ [ linux-4.17.19 ] ~ [ linux-4.16.18 ] ~ [ linux-4.15.18 ] ~ [ linux-4.14.336 ] ~ [ linux-4.13.16 ] ~ [ linux-4.12.14 ] ~ [ linux-4.11.12 ] ~ [ linux-4.10.17 ] ~ [ linux-4.9.337 ] ~ [ linux-4.4.302 ] ~ [ linux-3.10.108 ] ~ [ linux-2.6.32.71 ] ~ [ linux-2.6.0 ] ~ [ linux-2.4.37.11 ] ~ [ unix-v6-master ] ~ [ ccs-tools-1.8.12 ] ~ [ policy-sample ] ~
Architecture: ~ [ i386 ] ~ [ alpha ] ~ [ m68k ] ~ [ mips ] ~ [ ppc ] ~ [ sparc ] ~ [ sparc64 ] ~

  1 /* SPDX-License-Identifier: Apache-2.0 OR BSD-2-Clause */
  2 //
  3 // VAES and VPCLMULQDQ optimized AES-GCM for x86_64
  4 //
  5 // Copyright 2024 Google LLC
  6 //
  7 // Author: Eric Biggers <ebiggers@google.com>
  8 //
  9 //------------------------------------------------------------------------------
 10 //
 11 // This file is dual-licensed, meaning that you can use it under your choice of
 12 // either of the following two licenses:
 13 //
 14 // Licensed under the Apache License 2.0 (the "License").  You may obtain a copy
 15 // of the License at
 16 //
 17 //      http://www.apache.org/licenses/LICENSE-2.0
 18 //
 19 // Unless required by applicable law or agreed to in writing, software
 20 // distributed under the License is distributed on an "AS IS" BASIS,
 21 // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 22 // See the License for the specific language governing permissions and
 23 // limitations under the License.
 24 //
 25 // or
 26 //
 27 // Redistribution and use in source and binary forms, with or without
 28 // modification, are permitted provided that the following conditions are met:
 29 //
 30 // 1. Redistributions of source code must retain the above copyright notice,
 31 //    this list of conditions and the following disclaimer.
 32 //
 33 // 2. Redistributions in binary form must reproduce the above copyright
 34 //    notice, this list of conditions and the following disclaimer in the
 35 //    documentation and/or other materials provided with the distribution.
 36 //
 37 // THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
 38 // AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
 39 // IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
 40 // ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE
 41 // LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
 42 // CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
 43 // SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
 44 // INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
 45 // CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
 46 // ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
 47 // POSSIBILITY OF SUCH DAMAGE.
 48 //
 49 //------------------------------------------------------------------------------
 50 //
 51 // This file implements AES-GCM (Galois/Counter Mode) for x86_64 CPUs that
 52 // support VAES (vector AES), VPCLMULQDQ (vector carryless multiplication), and
 53 // either AVX512 or AVX10.  Some of the functions, notably the encryption and
 54 // decryption update functions which are the most performance-critical, are
 55 // provided in two variants generated from a macro: one using 256-bit vectors
 56 // (suffix: vaes_avx10_256) and one using 512-bit vectors (vaes_avx10_512).  The
 57 // other, "shared" functions (vaes_avx10) use at most 256-bit vectors.
 58 //
 59 // The functions that use 512-bit vectors are intended for CPUs that support
 60 // 512-bit vectors *and* where using them doesn't cause significant
 61 // downclocking.  They require the following CPU features:
 62 //
 63 //      VAES && VPCLMULQDQ && BMI2 && ((AVX512BW && AVX512VL) || AVX10/512)
 64 //
 65 // The other functions require the following CPU features:
 66 //
 67 //      VAES && VPCLMULQDQ && BMI2 && ((AVX512BW && AVX512VL) || AVX10/256)
 68 //
 69 // All functions use the "System V" ABI.  The Windows ABI is not supported.
 70 //
 71 // Note that we use "avx10" in the names of the functions as a shorthand to
 72 // really mean "AVX10 or a certain set of AVX512 features".  Due to Intel's
 73 // introduction of AVX512 and then its replacement by AVX10, there doesn't seem
 74 // to be a simple way to name things that makes sense on all CPUs.
 75 //
 76 // Note that the macros that support both 256-bit and 512-bit vectors could
 77 // fairly easily be changed to support 128-bit too.  However, this would *not*
 78 // be sufficient to allow the code to run on CPUs without AVX512 or AVX10,
 79 // because the code heavily uses several features of these extensions other than
 80 // the vector length: the increase in the number of SIMD registers from 16 to
 81 // 32, masking support, and new instructions such as vpternlogd (which can do a
 82 // three-argument XOR).  These features are very useful for AES-GCM.
 83 
 84 #include <linux/linkage.h>
 85 
 86 .section .rodata
 87 .p2align 6
 88 
 89         // A shuffle mask that reflects the bytes of 16-byte blocks
 90 .Lbswap_mask:
 91         .octa   0x000102030405060708090a0b0c0d0e0f
 92 
 93         // This is the GHASH reducing polynomial without its constant term, i.e.
 94         // x^128 + x^7 + x^2 + x, represented using the backwards mapping
 95         // between bits and polynomial coefficients.
 96         //
 97         // Alternatively, it can be interpreted as the naturally-ordered
 98         // representation of the polynomial x^127 + x^126 + x^121 + 1, i.e. the
 99         // "reversed" GHASH reducing polynomial without its x^128 term.
100 .Lgfpoly:
101         .octa   0xc2000000000000000000000000000001
102 
103         // Same as above, but with the (1 << 64) bit set.
104 .Lgfpoly_and_internal_carrybit:
105         .octa   0xc2000000000000010000000000000001
106 
107         // The below constants are used for incrementing the counter blocks.
108         // ctr_pattern points to the four 128-bit values [0, 1, 2, 3].
109         // inc_2blocks and inc_4blocks point to the single 128-bit values 2 and
110         // 4.  Note that the same '2' is reused in ctr_pattern and inc_2blocks.
111 .Lctr_pattern:
112         .octa   0
113         .octa   1
114 .Linc_2blocks:
115         .octa   2
116         .octa   3
117 .Linc_4blocks:
118         .octa   4
119 
120 // Number of powers of the hash key stored in the key struct.  The powers are
121 // stored from highest (H^NUM_H_POWERS) to lowest (H^1).
122 #define NUM_H_POWERS            16
123 
124 // Offset to AES key length (in bytes) in the key struct
125 #define OFFSETOF_AESKEYLEN      480
126 
127 // Offset to start of hash key powers array in the key struct
128 #define OFFSETOF_H_POWERS       512
129 
130 // Offset to end of hash key powers array in the key struct.
131 //
132 // This is immediately followed by three zeroized padding blocks, which are
133 // included so that partial vectors can be handled more easily.  E.g. if VL=64
134 // and two blocks remain, we load the 4 values [H^2, H^1, 0, 0].  The most
135 // padding blocks needed is 3, which occurs if [H^1, 0, 0, 0] is loaded.
136 #define OFFSETOFEND_H_POWERS    (OFFSETOF_H_POWERS + (NUM_H_POWERS * 16))
137 
138 .text
139 
140 // Set the vector length in bytes.  This sets the VL variable and defines
141 // register aliases V0-V31 that map to the ymm or zmm registers.
142 .macro  _set_veclen     vl
143         .set    VL,     \vl
144 .irp i, 0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15, \
145         16,17,18,19,20,21,22,23,24,25,26,27,28,29,30,31
146 .if VL == 32
147         .set    V\i,    %ymm\i
148 .elseif VL == 64
149         .set    V\i,    %zmm\i
150 .else
151         .error "Unsupported vector length"
152 .endif
153 .endr
154 .endm
155 
156 // The _ghash_mul_step macro does one step of GHASH multiplication of the
157 // 128-bit lanes of \a by the corresponding 128-bit lanes of \b and storing the
158 // reduced products in \dst.  \t0, \t1, and \t2 are temporary registers of the
159 // same size as \a and \b.  To complete all steps, this must invoked with \i=0
160 // through \i=9.  The division into steps allows users of this macro to
161 // optionally interleave the computation with other instructions.  Users of this
162 // macro must preserve the parameter registers across steps.
163 //
164 // The multiplications are done in GHASH's representation of the finite field
165 // GF(2^128).  Elements of GF(2^128) are represented as binary polynomials
166 // (i.e. polynomials whose coefficients are bits) modulo a reducing polynomial
167 // G.  The GCM specification uses G = x^128 + x^7 + x^2 + x + 1.  Addition is
168 // just XOR, while multiplication is more complex and has two parts: (a) do
169 // carryless multiplication of two 128-bit input polynomials to get a 256-bit
170 // intermediate product polynomial, and (b) reduce the intermediate product to
171 // 128 bits by adding multiples of G that cancel out terms in it.  (Adding
172 // multiples of G doesn't change which field element the polynomial represents.)
173 //
174 // Unfortunately, the GCM specification maps bits to/from polynomial
175 // coefficients backwards from the natural order.  In each byte it specifies the
176 // highest bit to be the lowest order polynomial coefficient, *not* the highest!
177 // This makes it nontrivial to work with the GHASH polynomials.  We could
178 // reflect the bits, but x86 doesn't have an instruction that does that.
179 //
180 // Instead, we operate on the values without bit-reflecting them.  This *mostly*
181 // just works, since XOR and carryless multiplication are symmetric with respect
182 // to bit order, but it has some consequences.  First, due to GHASH's byte
183 // order, by skipping bit reflection, *byte* reflection becomes necessary to
184 // give the polynomial terms a consistent order.  E.g., considering an N-bit
185 // value interpreted using the G = x^128 + x^7 + x^2 + x + 1 convention, bits 0
186 // through N-1 of the byte-reflected value represent the coefficients of x^(N-1)
187 // through x^0, whereas bits 0 through N-1 of the non-byte-reflected value
188 // represent x^7...x^0, x^15...x^8, ..., x^(N-1)...x^(N-8) which can't be worked
189 // with.  Fortunately, x86's vpshufb instruction can do byte reflection.
190 //
191 // Second, forgoing the bit reflection causes an extra multiple of x (still
192 // using the G = x^128 + x^7 + x^2 + x + 1 convention) to be introduced by each
193 // multiplication.  This is because an M-bit by N-bit carryless multiplication
194 // really produces a (M+N-1)-bit product, but in practice it's zero-extended to
195 // M+N bits.  In the G = x^128 + x^7 + x^2 + x + 1 convention, which maps bits
196 // to polynomial coefficients backwards, this zero-extension actually changes
197 // the product by introducing an extra factor of x.  Therefore, users of this
198 // macro must ensure that one of the inputs has an extra factor of x^-1, i.e.
199 // the multiplicative inverse of x, to cancel out the extra x.
200 //
201 // Third, the backwards coefficients convention is just confusing to work with,
202 // since it makes "low" and "high" in the polynomial math mean the opposite of
203 // their normal meaning in computer programming.  This can be solved by using an
204 // alternative interpretation: the polynomial coefficients are understood to be
205 // in the natural order, and the multiplication is actually \a * \b * x^-128 mod
206 // x^128 + x^127 + x^126 + x^121 + 1.  This doesn't change the inputs, outputs,
207 // or the implementation at all; it just changes the mathematical interpretation
208 // of what each instruction is doing.  Starting from here, we'll use this
209 // alternative interpretation, as it's easier to understand the code that way.
210 //
211 // Moving onto the implementation, the vpclmulqdq instruction does 64 x 64 =>
212 // 128-bit carryless multiplication, so we break the 128 x 128 multiplication
213 // into parts as follows (the _L and _H suffixes denote low and high 64 bits):
214 //
215 //     LO = a_L * b_L
216 //     MI = (a_L * b_H) + (a_H * b_L)
217 //     HI = a_H * b_H
218 //
219 // The 256-bit product is x^128*HI + x^64*MI + LO.  LO, MI, and HI are 128-bit.
220 // Note that MI "overlaps" with LO and HI.  We don't consolidate MI into LO and
221 // HI right away, since the way the reduction works makes that unnecessary.
222 //
223 // For the reduction, we cancel out the low 128 bits by adding multiples of G =
224 // x^128 + x^127 + x^126 + x^121 + 1.  This is done by two iterations, each of
225 // which cancels out the next lowest 64 bits.  Consider a value x^64*A + B,
226 // where A and B are 128-bit.  Adding B_L*G to that value gives:
227 //
228 //       x^64*A + B + B_L*G
229 //     = x^64*A + x^64*B_H + B_L + B_L*(x^128 + x^127 + x^126 + x^121 + 1)
230 //     = x^64*A + x^64*B_H + B_L + x^128*B_L + x^64*B_L*(x^63 + x^62 + x^57) + B_L
231 //     = x^64*A + x^64*B_H + x^128*B_L + x^64*B_L*(x^63 + x^62 + x^57) + B_L + B_L
232 //     = x^64*(A + B_H + x^64*B_L + B_L*(x^63 + x^62 + x^57))
233 //
234 // So: if we sum A, B with its halves swapped, and the low half of B times x^63
235 // + x^62 + x^57, we get a 128-bit value C where x^64*C is congruent to the
236 // original value x^64*A + B.  I.e., the low 64 bits got canceled out.
237 //
238 // We just need to apply this twice: first to fold LO into MI, and second to
239 // fold the updated MI into HI.
240 //
241 // The needed three-argument XORs are done using the vpternlogd instruction with
242 // immediate 0x96, since this is faster than two vpxord instructions.
243 //
244 // A potential optimization, assuming that b is fixed per-key (if a is fixed
245 // per-key it would work the other way around), is to use one iteration of the
246 // reduction described above to precompute a value c such that x^64*c = b mod G,
247 // and then multiply a_L by c (and implicitly by x^64) instead of by b:
248 //
249 //     MI = (a_L * c_L) + (a_H * b_L)
250 //     HI = (a_L * c_H) + (a_H * b_H)
251 //
252 // This would eliminate the LO part of the intermediate product, which would
253 // eliminate the need to fold LO into MI.  This would save two instructions,
254 // including a vpclmulqdq.  However, we currently don't use this optimization
255 // because it would require twice as many per-key precomputed values.
256 //
257 // Using Karatsuba multiplication instead of "schoolbook" multiplication
258 // similarly would save a vpclmulqdq but does not seem to be worth it.
259 .macro  _ghash_mul_step i, a, b, dst, gfpoly, t0, t1, t2
260 .if \i == 0
261         vpclmulqdq      $0x00, \a, \b, \t0        // LO = a_L * b_L
262         vpclmulqdq      $0x01, \a, \b, \t1        // MI_0 = a_L * b_H
263 .elseif \i == 1
264         vpclmulqdq      $0x10, \a, \b, \t2        // MI_1 = a_H * b_L
265 .elseif \i == 2
266         vpxord          \t2, \t1, \t1             // MI = MI_0 + MI_1
267 .elseif \i == 3
268         vpclmulqdq      $0x01, \t0, \gfpoly, \t2  // LO_L*(x^63 + x^62 + x^57)
269 .elseif \i == 4
270         vpshufd         $0x4e, \t0, \t0           // Swap halves of LO
271 .elseif \i == 5
272         vpternlogd      $0x96, \t2, \t0, \t1      // Fold LO into MI
273 .elseif \i == 6
274         vpclmulqdq      $0x11, \a, \b, \dst       // HI = a_H * b_H
275 .elseif \i == 7
276         vpclmulqdq      $0x01, \t1, \gfpoly, \t0  // MI_L*(x^63 + x^62 + x^57)
277 .elseif \i == 8
278         vpshufd         $0x4e, \t1, \t1           // Swap halves of MI
279 .elseif \i == 9
280         vpternlogd      $0x96, \t0, \t1, \dst     // Fold MI into HI
281 .endif
282 .endm
283 
284 // GHASH-multiply the 128-bit lanes of \a by the 128-bit lanes of \b and store
285 // the reduced products in \dst.  See _ghash_mul_step for full explanation.
286 .macro  _ghash_mul      a, b, dst, gfpoly, t0, t1, t2
287 .irp i, 0,1,2,3,4,5,6,7,8,9
288         _ghash_mul_step \i, \a, \b, \dst, \gfpoly, \t0, \t1, \t2
289 .endr
290 .endm
291 
292 // GHASH-multiply the 128-bit lanes of \a by the 128-bit lanes of \b and add the
293 // *unreduced* products to \lo, \mi, and \hi.
294 .macro  _ghash_mul_noreduce     a, b, lo, mi, hi, t0, t1, t2, t3
295         vpclmulqdq      $0x00, \a, \b, \t0      // a_L * b_L
296         vpclmulqdq      $0x01, \a, \b, \t1      // a_L * b_H
297         vpclmulqdq      $0x10, \a, \b, \t2      // a_H * b_L
298         vpclmulqdq      $0x11, \a, \b, \t3      // a_H * b_H
299         vpxord          \t0, \lo, \lo
300         vpternlogd      $0x96, \t2, \t1, \mi
301         vpxord          \t3, \hi, \hi
302 .endm
303 
304 // Reduce the unreduced products from \lo, \mi, and \hi and store the 128-bit
305 // reduced products in \hi.  See _ghash_mul_step for explanation of reduction.
306 .macro  _ghash_reduce   lo, mi, hi, gfpoly, t0
307         vpclmulqdq      $0x01, \lo, \gfpoly, \t0
308         vpshufd         $0x4e, \lo, \lo
309         vpternlogd      $0x96, \t0, \lo, \mi
310         vpclmulqdq      $0x01, \mi, \gfpoly, \t0
311         vpshufd         $0x4e, \mi, \mi
312         vpternlogd      $0x96, \t0, \mi, \hi
313 .endm
314 
315 // void aes_gcm_precompute_##suffix(struct aes_gcm_key_avx10 *key);
316 //
317 // Given the expanded AES key |key->aes_key|, this function derives the GHASH
318 // subkey and initializes |key->ghash_key_powers| with powers of it.
319 //
320 // The number of key powers initialized is NUM_H_POWERS, and they are stored in
321 // the order H^NUM_H_POWERS to H^1.  The zeroized padding blocks after the key
322 // powers themselves are also initialized.
323 //
324 // This macro supports both VL=32 and VL=64.  _set_veclen must have been invoked
325 // with the desired length.  In the VL=32 case, the function computes twice as
326 // many key powers than are actually used by the VL=32 GCM update functions.
327 // This is done to keep the key format the same regardless of vector length.
328 .macro  _aes_gcm_precompute
329 
330         // Function arguments
331         .set    KEY,            %rdi
332 
333         // Additional local variables.  V0-V2 and %rax are used as temporaries.
334         .set    POWERS_PTR,     %rsi
335         .set    RNDKEYLAST_PTR, %rdx
336         .set    H_CUR,          V3
337         .set    H_CUR_YMM,      %ymm3
338         .set    H_CUR_XMM,      %xmm3
339         .set    H_INC,          V4
340         .set    H_INC_YMM,      %ymm4
341         .set    H_INC_XMM,      %xmm4
342         .set    GFPOLY,         V5
343         .set    GFPOLY_YMM,     %ymm5
344         .set    GFPOLY_XMM,     %xmm5
345 
346         // Get pointer to lowest set of key powers (located at end of array).
347         lea             OFFSETOFEND_H_POWERS-VL(KEY), POWERS_PTR
348 
349         // Encrypt an all-zeroes block to get the raw hash subkey.
350         movl            OFFSETOF_AESKEYLEN(KEY), %eax
351         lea             6*16(KEY,%rax,4), RNDKEYLAST_PTR
352         vmovdqu         (KEY), %xmm0  // Zero-th round key XOR all-zeroes block
353         add             $16, KEY
354 1:
355         vaesenc         (KEY), %xmm0, %xmm0
356         add             $16, KEY
357         cmp             KEY, RNDKEYLAST_PTR
358         jne             1b
359         vaesenclast     (RNDKEYLAST_PTR), %xmm0, %xmm0
360 
361         // Reflect the bytes of the raw hash subkey.
362         vpshufb         .Lbswap_mask(%rip), %xmm0, H_CUR_XMM
363 
364         // Zeroize the padding blocks.
365         vpxor           %xmm0, %xmm0, %xmm0
366         vmovdqu         %ymm0, VL(POWERS_PTR)
367         vmovdqu         %xmm0, VL+2*16(POWERS_PTR)
368 
369         // Finish preprocessing the first key power, H^1.  Since this GHASH
370         // implementation operates directly on values with the backwards bit
371         // order specified by the GCM standard, it's necessary to preprocess the
372         // raw key as follows.  First, reflect its bytes.  Second, multiply it
373         // by x^-1 mod x^128 + x^7 + x^2 + x + 1 (if using the backwards
374         // interpretation of polynomial coefficients), which can also be
375         // interpreted as multiplication by x mod x^128 + x^127 + x^126 + x^121
376         // + 1 using the alternative, natural interpretation of polynomial
377         // coefficients.  For details, see the comment above _ghash_mul_step.
378         //
379         // Either way, for the multiplication the concrete operation performed
380         // is a left shift of the 128-bit value by 1 bit, then an XOR with (0xc2
381         // << 120) | 1 if a 1 bit was carried out.  However, there's no 128-bit
382         // wide shift instruction, so instead double each of the two 64-bit
383         // halves and incorporate the internal carry bit into the value XOR'd.
384         vpshufd         $0xd3, H_CUR_XMM, %xmm0
385         vpsrad          $31, %xmm0, %xmm0
386         vpaddq          H_CUR_XMM, H_CUR_XMM, H_CUR_XMM
387         vpand           .Lgfpoly_and_internal_carrybit(%rip), %xmm0, %xmm0
388         vpxor           %xmm0, H_CUR_XMM, H_CUR_XMM
389 
390         // Load the gfpoly constant.
391         vbroadcasti32x4 .Lgfpoly(%rip), GFPOLY
392 
393         // Square H^1 to get H^2.
394         //
395         // Note that as with H^1, all higher key powers also need an extra
396         // factor of x^-1 (or x using the natural interpretation).  Nothing
397         // special needs to be done to make this happen, though: H^1 * H^1 would
398         // end up with two factors of x^-1, but the multiplication consumes one.
399         // So the product H^2 ends up with the desired one factor of x^-1.
400         _ghash_mul      H_CUR_XMM, H_CUR_XMM, H_INC_XMM, GFPOLY_XMM, \
401                         %xmm0, %xmm1, %xmm2
402 
403         // Create H_CUR_YMM = [H^2, H^1] and H_INC_YMM = [H^2, H^2].
404         vinserti128     $1, H_CUR_XMM, H_INC_YMM, H_CUR_YMM
405         vinserti128     $1, H_INC_XMM, H_INC_YMM, H_INC_YMM
406 
407 .if VL == 64
408         // Create H_CUR = [H^4, H^3, H^2, H^1] and H_INC = [H^4, H^4, H^4, H^4].
409         _ghash_mul      H_INC_YMM, H_CUR_YMM, H_INC_YMM, GFPOLY_YMM, \
410                         %ymm0, %ymm1, %ymm2
411         vinserti64x4    $1, H_CUR_YMM, H_INC, H_CUR
412         vshufi64x2      $0, H_INC, H_INC, H_INC
413 .endif
414 
415         // Store the lowest set of key powers.
416         vmovdqu8        H_CUR, (POWERS_PTR)
417 
418         // Compute and store the remaining key powers.  With VL=32, repeatedly
419         // multiply [H^(i+1), H^i] by [H^2, H^2] to get [H^(i+3), H^(i+2)].
420         // With VL=64, repeatedly multiply [H^(i+3), H^(i+2), H^(i+1), H^i] by
421         // [H^4, H^4, H^4, H^4] to get [H^(i+7), H^(i+6), H^(i+5), H^(i+4)].
422         mov             $(NUM_H_POWERS*16/VL) - 1, %eax
423 .Lprecompute_next\@:
424         sub             $VL, POWERS_PTR
425         _ghash_mul      H_INC, H_CUR, H_CUR, GFPOLY, V0, V1, V2
426         vmovdqu8        H_CUR, (POWERS_PTR)
427         dec             %eax
428         jnz             .Lprecompute_next\@
429 
430         vzeroupper      // This is needed after using ymm or zmm registers.
431         RET
432 .endm
433 
434 // XOR together the 128-bit lanes of \src (whose low lane is \src_xmm) and store
435 // the result in \dst_xmm.  This implicitly zeroizes the other lanes of dst.
436 .macro  _horizontal_xor src, src_xmm, dst_xmm, t0_xmm, t1_xmm, t2_xmm
437         vextracti32x4   $1, \src, \t0_xmm
438 .if VL == 32
439         vpxord          \t0_xmm, \src_xmm, \dst_xmm
440 .elseif VL == 64
441         vextracti32x4   $2, \src, \t1_xmm
442         vextracti32x4   $3, \src, \t2_xmm
443         vpxord          \t0_xmm, \src_xmm, \dst_xmm
444         vpternlogd      $0x96, \t1_xmm, \t2_xmm, \dst_xmm
445 .else
446         .error "Unsupported vector length"
447 .endif
448 .endm
449 
450 // Do one step of the GHASH update of the data blocks given in the vector
451 // registers GHASHDATA[0-3].  \i specifies the step to do, 0 through 9.  The
452 // division into steps allows users of this macro to optionally interleave the
453 // computation with other instructions.  This macro uses the vector register
454 // GHASH_ACC as input/output; GHASHDATA[0-3] as inputs that are clobbered;
455 // H_POW[4-1], GFPOLY, and BSWAP_MASK as inputs that aren't clobbered; and
456 // GHASHTMP[0-2] as temporaries.  This macro handles the byte-reflection of the
457 // data blocks.  The parameter registers must be preserved across steps.
458 //
459 // The GHASH update does: GHASH_ACC = H_POW4*(GHASHDATA0 + GHASH_ACC) +
460 // H_POW3*GHASHDATA1 + H_POW2*GHASHDATA2 + H_POW1*GHASHDATA3, where the
461 // operations are vectorized operations on vectors of 16-byte blocks.  E.g.,
462 // with VL=32 there are 2 blocks per vector and the vectorized terms correspond
463 // to the following non-vectorized terms:
464 //
465 //      H_POW4*(GHASHDATA0 + GHASH_ACC) => H^8*(blk0 + GHASH_ACC_XMM) and H^7*(blk1 + 0)
466 //      H_POW3*GHASHDATA1 => H^6*blk2 and H^5*blk3
467 //      H_POW2*GHASHDATA2 => H^4*blk4 and H^3*blk5
468 //      H_POW1*GHASHDATA3 => H^2*blk6 and H^1*blk7
469 //
470 // With VL=64, we use 4 blocks/vector, H^16 through H^1, and blk0 through blk15.
471 //
472 // More concretely, this code does:
473 //   - Do vectorized "schoolbook" multiplications to compute the intermediate
474 //     256-bit product of each block and its corresponding hash key power.
475 //     There are 4*VL/16 of these intermediate products.
476 //   - Sum (XOR) the intermediate 256-bit products across vectors.  This leaves
477 //     VL/16 256-bit intermediate values.
478 //   - Do a vectorized reduction of these 256-bit intermediate values to
479 //     128-bits each.  This leaves VL/16 128-bit intermediate values.
480 //   - Sum (XOR) these values and store the 128-bit result in GHASH_ACC_XMM.
481 //
482 // See _ghash_mul_step for the full explanation of the operations performed for
483 // each individual finite field multiplication and reduction.
484 .macro  _ghash_step_4x  i
485 .if \i == 0
486         vpshufb         BSWAP_MASK, GHASHDATA0, GHASHDATA0
487         vpxord          GHASH_ACC, GHASHDATA0, GHASHDATA0
488         vpshufb         BSWAP_MASK, GHASHDATA1, GHASHDATA1
489         vpshufb         BSWAP_MASK, GHASHDATA2, GHASHDATA2
490 .elseif \i == 1
491         vpshufb         BSWAP_MASK, GHASHDATA3, GHASHDATA3
492         vpclmulqdq      $0x00, H_POW4, GHASHDATA0, GHASH_ACC    // LO_0
493         vpclmulqdq      $0x00, H_POW3, GHASHDATA1, GHASHTMP0    // LO_1
494         vpclmulqdq      $0x00, H_POW2, GHASHDATA2, GHASHTMP1    // LO_2
495 .elseif \i == 2
496         vpxord          GHASHTMP0, GHASH_ACC, GHASH_ACC         // sum(LO_{1,0})
497         vpclmulqdq      $0x00, H_POW1, GHASHDATA3, GHASHTMP2    // LO_3
498         vpternlogd      $0x96, GHASHTMP2, GHASHTMP1, GHASH_ACC  // LO = sum(LO_{3,2,1,0})
499         vpclmulqdq      $0x01, H_POW4, GHASHDATA0, GHASHTMP0    // MI_0
500 .elseif \i == 3
501         vpclmulqdq      $0x01, H_POW3, GHASHDATA1, GHASHTMP1    // MI_1
502         vpclmulqdq      $0x01, H_POW2, GHASHDATA2, GHASHTMP2    // MI_2
503         vpternlogd      $0x96, GHASHTMP2, GHASHTMP1, GHASHTMP0  // sum(MI_{2,1,0})
504         vpclmulqdq      $0x01, H_POW1, GHASHDATA3, GHASHTMP1    // MI_3
505 .elseif \i == 4
506         vpclmulqdq      $0x10, H_POW4, GHASHDATA0, GHASHTMP2    // MI_4
507         vpternlogd      $0x96, GHASHTMP2, GHASHTMP1, GHASHTMP0  // sum(MI_{4,3,2,1,0})
508         vpclmulqdq      $0x10, H_POW3, GHASHDATA1, GHASHTMP1    // MI_5
509         vpclmulqdq      $0x10, H_POW2, GHASHDATA2, GHASHTMP2    // MI_6
510 .elseif \i == 5
511         vpternlogd      $0x96, GHASHTMP2, GHASHTMP1, GHASHTMP0  // sum(MI_{6,5,4,3,2,1,0})
512         vpclmulqdq      $0x01, GHASH_ACC, GFPOLY, GHASHTMP2     // LO_L*(x^63 + x^62 + x^57)
513         vpclmulqdq      $0x10, H_POW1, GHASHDATA3, GHASHTMP1    // MI_7
514         vpxord          GHASHTMP1, GHASHTMP0, GHASHTMP0         // MI = sum(MI_{7,6,5,4,3,2,1,0})
515 .elseif \i == 6
516         vpshufd         $0x4e, GHASH_ACC, GHASH_ACC             // Swap halves of LO
517         vpclmulqdq      $0x11, H_POW4, GHASHDATA0, GHASHDATA0   // HI_0
518         vpclmulqdq      $0x11, H_POW3, GHASHDATA1, GHASHDATA1   // HI_1
519         vpclmulqdq      $0x11, H_POW2, GHASHDATA2, GHASHDATA2   // HI_2
520 .elseif \i == 7
521         vpternlogd      $0x96, GHASHTMP2, GHASH_ACC, GHASHTMP0  // Fold LO into MI
522         vpclmulqdq      $0x11, H_POW1, GHASHDATA3, GHASHDATA3   // HI_3
523         vpternlogd      $0x96, GHASHDATA2, GHASHDATA1, GHASHDATA0 // sum(HI_{2,1,0})
524         vpclmulqdq      $0x01, GHASHTMP0, GFPOLY, GHASHTMP1     // MI_L*(x^63 + x^62 + x^57)
525 .elseif \i == 8
526         vpxord          GHASHDATA3, GHASHDATA0, GHASH_ACC       // HI = sum(HI_{3,2,1,0})
527         vpshufd         $0x4e, GHASHTMP0, GHASHTMP0             // Swap halves of MI
528         vpternlogd      $0x96, GHASHTMP1, GHASHTMP0, GHASH_ACC  // Fold MI into HI
529 .elseif \i == 9
530         _horizontal_xor GHASH_ACC, GHASH_ACC_XMM, GHASH_ACC_XMM, \
531                         GHASHDATA0_XMM, GHASHDATA1_XMM, GHASHDATA2_XMM
532 .endif
533 .endm
534 
535 // Do one non-last round of AES encryption on the counter blocks in V0-V3 using
536 // the round key that has been broadcast to all 128-bit lanes of \round_key.
537 .macro  _vaesenc_4x     round_key
538         vaesenc         \round_key, V0, V0
539         vaesenc         \round_key, V1, V1
540         vaesenc         \round_key, V2, V2
541         vaesenc         \round_key, V3, V3
542 .endm
543 
544 // Start the AES encryption of four vectors of counter blocks.
545 .macro  _ctr_begin_4x
546 
547         // Increment LE_CTR four times to generate four vectors of little-endian
548         // counter blocks, swap each to big-endian, and store them in V0-V3.
549         vpshufb         BSWAP_MASK, LE_CTR, V0
550         vpaddd          LE_CTR_INC, LE_CTR, LE_CTR
551         vpshufb         BSWAP_MASK, LE_CTR, V1
552         vpaddd          LE_CTR_INC, LE_CTR, LE_CTR
553         vpshufb         BSWAP_MASK, LE_CTR, V2
554         vpaddd          LE_CTR_INC, LE_CTR, LE_CTR
555         vpshufb         BSWAP_MASK, LE_CTR, V3
556         vpaddd          LE_CTR_INC, LE_CTR, LE_CTR
557 
558         // AES "round zero": XOR in the zero-th round key.
559         vpxord          RNDKEY0, V0, V0
560         vpxord          RNDKEY0, V1, V1
561         vpxord          RNDKEY0, V2, V2
562         vpxord          RNDKEY0, V3, V3
563 .endm
564 
565 // void aes_gcm_{enc,dec}_update_##suffix(const struct aes_gcm_key_avx10 *key,
566 //                                        const u32 le_ctr[4], u8 ghash_acc[16],
567 //                                        const u8 *src, u8 *dst, int datalen);
568 //
569 // This macro generates a GCM encryption or decryption update function with the
570 // above prototype (with \enc selecting which one).  This macro supports both
571 // VL=32 and VL=64.  _set_veclen must have been invoked with the desired length.
572 //
573 // This function computes the next portion of the CTR keystream, XOR's it with
574 // |datalen| bytes from |src|, and writes the resulting encrypted or decrypted
575 // data to |dst|.  It also updates the GHASH accumulator |ghash_acc| using the
576 // next |datalen| ciphertext bytes.
577 //
578 // |datalen| must be a multiple of 16, except on the last call where it can be
579 // any length.  The caller must do any buffering needed to ensure this.  Both
580 // in-place and out-of-place en/decryption are supported.
581 //
582 // |le_ctr| must give the current counter in little-endian format.  For a new
583 // message, the low word of the counter must be 2.  This function loads the
584 // counter from |le_ctr| and increments the loaded counter as needed, but it
585 // does *not* store the updated counter back to |le_ctr|.  The caller must
586 // update |le_ctr| if any more data segments follow.  Internally, only the low
587 // 32-bit word of the counter is incremented, following the GCM standard.
588 .macro  _aes_gcm_update enc
589 
590         // Function arguments
591         .set    KEY,            %rdi
592         .set    LE_CTR_PTR,     %rsi
593         .set    GHASH_ACC_PTR,  %rdx
594         .set    SRC,            %rcx
595         .set    DST,            %r8
596         .set    DATALEN,        %r9d
597         .set    DATALEN64,      %r9     // Zero-extend DATALEN before using!
598 
599         // Additional local variables
600 
601         // %rax and %k1 are used as temporary registers.  LE_CTR_PTR is also
602         // available as a temporary register after the counter is loaded.
603 
604         // AES key length in bytes
605         .set    AESKEYLEN,      %r10d
606         .set    AESKEYLEN64,    %r10
607 
608         // Pointer to the last AES round key for the chosen AES variant
609         .set    RNDKEYLAST_PTR, %r11
610 
611         // In the main loop, V0-V3 are used as AES input and output.  Elsewhere
612         // they are used as temporary registers.
613 
614         // GHASHDATA[0-3] hold the ciphertext blocks and GHASH input data.
615         .set    GHASHDATA0,     V4
616         .set    GHASHDATA0_XMM, %xmm4
617         .set    GHASHDATA1,     V5
618         .set    GHASHDATA1_XMM, %xmm5
619         .set    GHASHDATA2,     V6
620         .set    GHASHDATA2_XMM, %xmm6
621         .set    GHASHDATA3,     V7
622 
623         // BSWAP_MASK is the shuffle mask for byte-reflecting 128-bit values
624         // using vpshufb, copied to all 128-bit lanes.
625         .set    BSWAP_MASK,     V8
626 
627         // RNDKEY temporarily holds the next AES round key.
628         .set    RNDKEY,         V9
629 
630         // GHASH_ACC is the accumulator variable for GHASH.  When fully reduced,
631         // only the lowest 128-bit lane can be nonzero.  When not fully reduced,
632         // more than one lane may be used, and they need to be XOR'd together.
633         .set    GHASH_ACC,      V10
634         .set    GHASH_ACC_XMM,  %xmm10
635 
636         // LE_CTR_INC is the vector of 32-bit words that need to be added to a
637         // vector of little-endian counter blocks to advance it forwards.
638         .set    LE_CTR_INC,     V11
639 
640         // LE_CTR contains the next set of little-endian counter blocks.
641         .set    LE_CTR,         V12
642 
643         // RNDKEY0, RNDKEYLAST, and RNDKEY_M[9-5] contain cached AES round keys,
644         // copied to all 128-bit lanes.  RNDKEY0 is the zero-th round key,
645         // RNDKEYLAST the last, and RNDKEY_M\i the one \i-th from the last.
646         .set    RNDKEY0,        V13
647         .set    RNDKEYLAST,     V14
648         .set    RNDKEY_M9,      V15
649         .set    RNDKEY_M8,      V16
650         .set    RNDKEY_M7,      V17
651         .set    RNDKEY_M6,      V18
652         .set    RNDKEY_M5,      V19
653 
654         // RNDKEYLAST[0-3] temporarily store the last AES round key XOR'd with
655         // the corresponding block of source data.  This is useful because
656         // vaesenclast(key, a) ^ b == vaesenclast(key ^ b, a), and key ^ b can
657         // be computed in parallel with the AES rounds.
658         .set    RNDKEYLAST0,    V20
659         .set    RNDKEYLAST1,    V21
660         .set    RNDKEYLAST2,    V22
661         .set    RNDKEYLAST3,    V23
662 
663         // GHASHTMP[0-2] are temporary variables used by _ghash_step_4x.  These
664         // cannot coincide with anything used for AES encryption, since for
665         // performance reasons GHASH and AES encryption are interleaved.
666         .set    GHASHTMP0,      V24
667         .set    GHASHTMP1,      V25
668         .set    GHASHTMP2,      V26
669 
670         // H_POW[4-1] contain the powers of the hash key H^(4*VL/16)...H^1.  The
671         // descending numbering reflects the order of the key powers.
672         .set    H_POW4,         V27
673         .set    H_POW3,         V28
674         .set    H_POW2,         V29
675         .set    H_POW1,         V30
676 
677         // GFPOLY contains the .Lgfpoly constant, copied to all 128-bit lanes.
678         .set    GFPOLY,         V31
679 
680         // Load some constants.
681         vbroadcasti32x4 .Lbswap_mask(%rip), BSWAP_MASK
682         vbroadcasti32x4 .Lgfpoly(%rip), GFPOLY
683 
684         // Load the GHASH accumulator and the starting counter.
685         vmovdqu         (GHASH_ACC_PTR), GHASH_ACC_XMM
686         vbroadcasti32x4 (LE_CTR_PTR), LE_CTR
687 
688         // Load the AES key length in bytes.
689         movl            OFFSETOF_AESKEYLEN(KEY), AESKEYLEN
690 
691         // Make RNDKEYLAST_PTR point to the last AES round key.  This is the
692         // round key with index 10, 12, or 14 for AES-128, AES-192, or AES-256
693         // respectively.  Then load the zero-th and last round keys.
694         lea             6*16(KEY,AESKEYLEN64,4), RNDKEYLAST_PTR
695         vbroadcasti32x4 (KEY), RNDKEY0
696         vbroadcasti32x4 (RNDKEYLAST_PTR), RNDKEYLAST
697 
698         // Finish initializing LE_CTR by adding [0, 1, ...] to its low words.
699         vpaddd          .Lctr_pattern(%rip), LE_CTR, LE_CTR
700 
701         // Initialize LE_CTR_INC to contain VL/16 in all 128-bit lanes.
702 .if VL == 32
703         vbroadcasti32x4 .Linc_2blocks(%rip), LE_CTR_INC
704 .elseif VL == 64
705         vbroadcasti32x4 .Linc_4blocks(%rip), LE_CTR_INC
706 .else
707         .error "Unsupported vector length"
708 .endif
709 
710         // If there are at least 4*VL bytes of data, then continue into the loop
711         // that processes 4*VL bytes of data at a time.  Otherwise skip it.
712         //
713         // Pre-subtracting 4*VL from DATALEN saves an instruction from the main
714         // loop and also ensures that at least one write always occurs to
715         // DATALEN, zero-extending it and allowing DATALEN64 to be used later.
716         sub             $4*VL, DATALEN
717         jl              .Lcrypt_loop_4x_done\@
718 
719         // Load powers of the hash key.
720         vmovdqu8        OFFSETOFEND_H_POWERS-4*VL(KEY), H_POW4
721         vmovdqu8        OFFSETOFEND_H_POWERS-3*VL(KEY), H_POW3
722         vmovdqu8        OFFSETOFEND_H_POWERS-2*VL(KEY), H_POW2
723         vmovdqu8        OFFSETOFEND_H_POWERS-1*VL(KEY), H_POW1
724 
725         // Main loop: en/decrypt and hash 4 vectors at a time.
726         //
727         // When possible, interleave the AES encryption of the counter blocks
728         // with the GHASH update of the ciphertext blocks.  This improves
729         // performance on many CPUs because the execution ports used by the VAES
730         // instructions often differ from those used by vpclmulqdq and other
731         // instructions used in GHASH.  For example, many Intel CPUs dispatch
732         // vaesenc to ports 0 and 1 and vpclmulqdq to port 5.
733         //
734         // The interleaving is easiest to do during decryption, since during
735         // decryption the ciphertext blocks are immediately available.  For
736         // encryption, instead encrypt the first set of blocks, then hash those
737         // blocks while encrypting the next set of blocks, repeat that as
738         // needed, and finally hash the last set of blocks.
739 
740 .if \enc
741         // Encrypt the first 4 vectors of plaintext blocks.  Leave the resulting
742         // ciphertext in GHASHDATA[0-3] for GHASH.
743         _ctr_begin_4x
744         lea             16(KEY), %rax
745 1:
746         vbroadcasti32x4 (%rax), RNDKEY
747         _vaesenc_4x     RNDKEY
748         add             $16, %rax
749         cmp             %rax, RNDKEYLAST_PTR
750         jne             1b
751         vpxord          0*VL(SRC), RNDKEYLAST, RNDKEYLAST0
752         vpxord          1*VL(SRC), RNDKEYLAST, RNDKEYLAST1
753         vpxord          2*VL(SRC), RNDKEYLAST, RNDKEYLAST2
754         vpxord          3*VL(SRC), RNDKEYLAST, RNDKEYLAST3
755         vaesenclast     RNDKEYLAST0, V0, GHASHDATA0
756         vaesenclast     RNDKEYLAST1, V1, GHASHDATA1
757         vaesenclast     RNDKEYLAST2, V2, GHASHDATA2
758         vaesenclast     RNDKEYLAST3, V3, GHASHDATA3
759         vmovdqu8        GHASHDATA0, 0*VL(DST)
760         vmovdqu8        GHASHDATA1, 1*VL(DST)
761         vmovdqu8        GHASHDATA2, 2*VL(DST)
762         vmovdqu8        GHASHDATA3, 3*VL(DST)
763         add             $4*VL, SRC
764         add             $4*VL, DST
765         sub             $4*VL, DATALEN
766         jl              .Lghash_last_ciphertext_4x\@
767 .endif
768 
769         // Cache as many additional AES round keys as possible.
770 .irp i, 9,8,7,6,5
771         vbroadcasti32x4 -\i*16(RNDKEYLAST_PTR), RNDKEY_M\i
772 .endr
773 
774 .Lcrypt_loop_4x\@:
775 
776         // If decrypting, load more ciphertext blocks into GHASHDATA[0-3].  If
777         // encrypting, GHASHDATA[0-3] already contain the previous ciphertext.
778 .if !\enc
779         vmovdqu8        0*VL(SRC), GHASHDATA0
780         vmovdqu8        1*VL(SRC), GHASHDATA1
781         vmovdqu8        2*VL(SRC), GHASHDATA2
782         vmovdqu8        3*VL(SRC), GHASHDATA3
783 .endif
784 
785         // Start the AES encryption of the counter blocks.
786         _ctr_begin_4x
787         cmp             $24, AESKEYLEN
788         jl              128f    // AES-128?
789         je              192f    // AES-192?
790         // AES-256
791         vbroadcasti32x4 -13*16(RNDKEYLAST_PTR), RNDKEY
792         _vaesenc_4x     RNDKEY
793         vbroadcasti32x4 -12*16(RNDKEYLAST_PTR), RNDKEY
794         _vaesenc_4x     RNDKEY
795 192:
796         vbroadcasti32x4 -11*16(RNDKEYLAST_PTR), RNDKEY
797         _vaesenc_4x     RNDKEY
798         vbroadcasti32x4 -10*16(RNDKEYLAST_PTR), RNDKEY
799         _vaesenc_4x     RNDKEY
800 128:
801 
802         // XOR the source data with the last round key, saving the result in
803         // RNDKEYLAST[0-3].  This reduces latency by taking advantage of the
804         // property vaesenclast(key, a) ^ b == vaesenclast(key ^ b, a).
805 .if \enc
806         vpxord          0*VL(SRC), RNDKEYLAST, RNDKEYLAST0
807         vpxord          1*VL(SRC), RNDKEYLAST, RNDKEYLAST1
808         vpxord          2*VL(SRC), RNDKEYLAST, RNDKEYLAST2
809         vpxord          3*VL(SRC), RNDKEYLAST, RNDKEYLAST3
810 .else
811         vpxord          GHASHDATA0, RNDKEYLAST, RNDKEYLAST0
812         vpxord          GHASHDATA1, RNDKEYLAST, RNDKEYLAST1
813         vpxord          GHASHDATA2, RNDKEYLAST, RNDKEYLAST2
814         vpxord          GHASHDATA3, RNDKEYLAST, RNDKEYLAST3
815 .endif
816 
817         // Finish the AES encryption of the counter blocks in V0-V3, interleaved
818         // with the GHASH update of the ciphertext blocks in GHASHDATA[0-3].
819 .irp i, 9,8,7,6,5
820         _vaesenc_4x     RNDKEY_M\i
821         _ghash_step_4x  (9 - \i)
822 .endr
823 .irp i, 4,3,2,1
824         vbroadcasti32x4 -\i*16(RNDKEYLAST_PTR), RNDKEY
825         _vaesenc_4x     RNDKEY
826         _ghash_step_4x  (9 - \i)
827 .endr
828         _ghash_step_4x  9
829 
830         // Do the last AES round.  This handles the XOR with the source data
831         // too, as per the optimization described above.
832         vaesenclast     RNDKEYLAST0, V0, GHASHDATA0
833         vaesenclast     RNDKEYLAST1, V1, GHASHDATA1
834         vaesenclast     RNDKEYLAST2, V2, GHASHDATA2
835         vaesenclast     RNDKEYLAST3, V3, GHASHDATA3
836 
837         // Store the en/decrypted data to DST.
838         vmovdqu8        GHASHDATA0, 0*VL(DST)
839         vmovdqu8        GHASHDATA1, 1*VL(DST)
840         vmovdqu8        GHASHDATA2, 2*VL(DST)
841         vmovdqu8        GHASHDATA3, 3*VL(DST)
842 
843         add             $4*VL, SRC
844         add             $4*VL, DST
845         sub             $4*VL, DATALEN
846         jge             .Lcrypt_loop_4x\@
847 
848 .if \enc
849 .Lghash_last_ciphertext_4x\@:
850         // Update GHASH with the last set of ciphertext blocks.
851 .irp i, 0,1,2,3,4,5,6,7,8,9
852         _ghash_step_4x  \i
853 .endr
854 .endif
855 
856 .Lcrypt_loop_4x_done\@:
857 
858         // Undo the extra subtraction by 4*VL and check whether data remains.
859         add             $4*VL, DATALEN
860         jz              .Ldone\@
861 
862         // The data length isn't a multiple of 4*VL.  Process the remaining data
863         // of length 1 <= DATALEN < 4*VL, up to one vector (VL bytes) at a time.
864         // Going one vector at a time may seem inefficient compared to having
865         // separate code paths for each possible number of vectors remaining.
866         // However, using a loop keeps the code size down, and it performs
867         // surprising well; modern CPUs will start executing the next iteration
868         // before the previous one finishes and also predict the number of loop
869         // iterations.  For a similar reason, we roll up the AES rounds.
870         //
871         // On the last iteration, the remaining length may be less than VL.
872         // Handle this using masking.
873         //
874         // Since there are enough key powers available for all remaining data,
875         // there is no need to do a GHASH reduction after each iteration.
876         // Instead, multiply each remaining block by its own key power, and only
877         // do a GHASH reduction at the very end.
878 
879         // Make POWERS_PTR point to the key powers [H^N, H^(N-1), ...] where N
880         // is the number of blocks that remain.
881         .set            POWERS_PTR, LE_CTR_PTR  // LE_CTR_PTR is free to be reused.
882         mov             DATALEN, %eax
883         neg             %rax
884         and             $~15, %rax  // -round_up(DATALEN, 16)
885         lea             OFFSETOFEND_H_POWERS(KEY,%rax), POWERS_PTR
886 
887         // Start collecting the unreduced GHASH intermediate value LO, MI, HI.
888         .set            LO, GHASHDATA0
889         .set            LO_XMM, GHASHDATA0_XMM
890         .set            MI, GHASHDATA1
891         .set            MI_XMM, GHASHDATA1_XMM
892         .set            HI, GHASHDATA2
893         .set            HI_XMM, GHASHDATA2_XMM
894         vpxor           LO_XMM, LO_XMM, LO_XMM
895         vpxor           MI_XMM, MI_XMM, MI_XMM
896         vpxor           HI_XMM, HI_XMM, HI_XMM
897 
898 .Lcrypt_loop_1x\@:
899 
900         // Select the appropriate mask for this iteration: all 1's if
901         // DATALEN >= VL, otherwise DATALEN 1's.  Do this branchlessly using the
902         // bzhi instruction from BMI2.  (This relies on DATALEN <= 255.)
903 .if VL < 64
904         mov             $-1, %eax
905         bzhi            DATALEN, %eax, %eax
906         kmovd           %eax, %k1
907 .else
908         mov             $-1, %rax
909         bzhi            DATALEN64, %rax, %rax
910         kmovq           %rax, %k1
911 .endif
912 
913         // Encrypt a vector of counter blocks.  This does not need to be masked.
914         vpshufb         BSWAP_MASK, LE_CTR, V0
915         vpaddd          LE_CTR_INC, LE_CTR, LE_CTR
916         vpxord          RNDKEY0, V0, V0
917         lea             16(KEY), %rax
918 1:
919         vbroadcasti32x4 (%rax), RNDKEY
920         vaesenc         RNDKEY, V0, V0
921         add             $16, %rax
922         cmp             %rax, RNDKEYLAST_PTR
923         jne             1b
924         vaesenclast     RNDKEYLAST, V0, V0
925 
926         // XOR the data with the appropriate number of keystream bytes.
927         vmovdqu8        (SRC), V1{%k1}{z}
928         vpxord          V1, V0, V0
929         vmovdqu8        V0, (DST){%k1}
930 
931         // Update GHASH with the ciphertext block(s), without reducing.
932         //
933         // In the case of DATALEN < VL, the ciphertext is zero-padded to VL.
934         // (If decrypting, it's done by the above masked load.  If encrypting,
935         // it's done by the below masked register-to-register move.)  Note that
936         // if DATALEN <= VL - 16, there will be additional padding beyond the
937         // padding of the last block specified by GHASH itself; i.e., there may
938         // be whole block(s) that get processed by the GHASH multiplication and
939         // reduction instructions but should not actually be included in the
940         // GHASH.  However, any such blocks are all-zeroes, and the values that
941         // they're multiplied with are also all-zeroes.  Therefore they just add
942         // 0 * 0 = 0 to the final GHASH result, which makes no difference.
943         vmovdqu8        (POWERS_PTR), H_POW1
944 .if \enc
945         vmovdqu8        V0, V1{%k1}{z}
946 .endif
947         vpshufb         BSWAP_MASK, V1, V0
948         vpxord          GHASH_ACC, V0, V0
949         _ghash_mul_noreduce     H_POW1, V0, LO, MI, HI, GHASHDATA3, V1, V2, V3
950         vpxor           GHASH_ACC_XMM, GHASH_ACC_XMM, GHASH_ACC_XMM
951 
952         add             $VL, POWERS_PTR
953         add             $VL, SRC
954         add             $VL, DST
955         sub             $VL, DATALEN
956         jg              .Lcrypt_loop_1x\@
957 
958         // Finally, do the GHASH reduction.
959         _ghash_reduce   LO, MI, HI, GFPOLY, V0
960         _horizontal_xor HI, HI_XMM, GHASH_ACC_XMM, %xmm0, %xmm1, %xmm2
961 
962 .Ldone\@:
963         // Store the updated GHASH accumulator back to memory.
964         vmovdqu         GHASH_ACC_XMM, (GHASH_ACC_PTR)
965 
966         vzeroupper      // This is needed after using ymm or zmm registers.
967         RET
968 .endm
969 
970 // void aes_gcm_enc_final_vaes_avx10(const struct aes_gcm_key_avx10 *key,
971 //                                   const u32 le_ctr[4], u8 ghash_acc[16],
972 //                                   u64 total_aadlen, u64 total_datalen);
973 // bool aes_gcm_dec_final_vaes_avx10(const struct aes_gcm_key_avx10 *key,
974 //                                   const u32 le_ctr[4],
975 //                                   const u8 ghash_acc[16],
976 //                                   u64 total_aadlen, u64 total_datalen,
977 //                                   const u8 tag[16], int taglen);
978 //
979 // This macro generates one of the above two functions (with \enc selecting
980 // which one).  Both functions finish computing the GCM authentication tag by
981 // updating GHASH with the lengths block and encrypting the GHASH accumulator.
982 // |total_aadlen| and |total_datalen| must be the total length of the additional
983 // authenticated data and the en/decrypted data in bytes, respectively.
984 //
985 // The encryption function then stores the full-length (16-byte) computed
986 // authentication tag to |ghash_acc|.  The decryption function instead loads the
987 // expected authentication tag (the one that was transmitted) from the 16-byte
988 // buffer |tag|, compares the first 4 <= |taglen| <= 16 bytes of it to the
989 // computed tag in constant time, and returns true if and only if they match.
990 .macro  _aes_gcm_final  enc
991 
992         // Function arguments
993         .set    KEY,            %rdi
994         .set    LE_CTR_PTR,     %rsi
995         .set    GHASH_ACC_PTR,  %rdx
996         .set    TOTAL_AADLEN,   %rcx
997         .set    TOTAL_DATALEN,  %r8
998         .set    TAG,            %r9
999         .set    TAGLEN,         %r10d   // Originally at 8(%rsp)
1000 
1001         // Additional local variables.
1002         // %rax, %xmm0-%xmm3, and %k1 are used as temporary registers.
1003         .set    AESKEYLEN,      %r11d
1004         .set    AESKEYLEN64,    %r11
1005         .set    GFPOLY,         %xmm4
1006         .set    BSWAP_MASK,     %xmm5
1007         .set    LE_CTR,         %xmm6
1008         .set    GHASH_ACC,      %xmm7
1009         .set    H_POW1,         %xmm8
1010 
1011         // Load some constants.
1012         vmovdqa         .Lgfpoly(%rip), GFPOLY
1013         vmovdqa         .Lbswap_mask(%rip), BSWAP_MASK
1014 
1015         // Load the AES key length in bytes.
1016         movl            OFFSETOF_AESKEYLEN(KEY), AESKEYLEN
1017 
1018         // Set up a counter block with 1 in the low 32-bit word.  This is the
1019         // counter that produces the ciphertext needed to encrypt the auth tag.
1020         // GFPOLY has 1 in the low word, so grab the 1 from there using a blend.
1021         vpblendd        $0xe, (LE_CTR_PTR), GFPOLY, LE_CTR
1022 
1023         // Build the lengths block and XOR it with the GHASH accumulator.
1024         // Although the lengths block is defined as the AAD length followed by
1025         // the en/decrypted data length, both in big-endian byte order, a byte
1026         // reflection of the full block is needed because of the way we compute
1027         // GHASH (see _ghash_mul_step).  By using little-endian values in the
1028         // opposite order, we avoid having to reflect any bytes here.
1029         vmovq           TOTAL_DATALEN, %xmm0
1030         vpinsrq         $1, TOTAL_AADLEN, %xmm0, %xmm0
1031         vpsllq          $3, %xmm0, %xmm0        // Bytes to bits
1032         vpxor           (GHASH_ACC_PTR), %xmm0, GHASH_ACC
1033 
1034         // Load the first hash key power (H^1), which is stored last.
1035         vmovdqu8        OFFSETOFEND_H_POWERS-16(KEY), H_POW1
1036 
1037 .if !\enc
1038         // Prepare a mask of TAGLEN one bits.
1039         movl            8(%rsp), TAGLEN
1040         mov             $-1, %eax
1041         bzhi            TAGLEN, %eax, %eax
1042         kmovd           %eax, %k1
1043 .endif
1044 
1045         // Make %rax point to the last AES round key for the chosen AES variant.
1046         lea             6*16(KEY,AESKEYLEN64,4), %rax
1047 
1048         // Start the AES encryption of the counter block by swapping the counter
1049         // block to big-endian and XOR-ing it with the zero-th AES round key.
1050         vpshufb         BSWAP_MASK, LE_CTR, %xmm0
1051         vpxor           (KEY), %xmm0, %xmm0
1052 
1053         // Complete the AES encryption and multiply GHASH_ACC by H^1.
1054         // Interleave the AES and GHASH instructions to improve performance.
1055         cmp             $24, AESKEYLEN
1056         jl              128f    // AES-128?
1057         je              192f    // AES-192?
1058         // AES-256
1059         vaesenc         -13*16(%rax), %xmm0, %xmm0
1060         vaesenc         -12*16(%rax), %xmm0, %xmm0
1061 192:
1062         vaesenc         -11*16(%rax), %xmm0, %xmm0
1063         vaesenc         -10*16(%rax), %xmm0, %xmm0
1064 128:
1065 .irp i, 0,1,2,3,4,5,6,7,8
1066         _ghash_mul_step \i, H_POW1, GHASH_ACC, GHASH_ACC, GFPOLY, \
1067                         %xmm1, %xmm2, %xmm3
1068         vaesenc         (\i-9)*16(%rax), %xmm0, %xmm0
1069 .endr
1070         _ghash_mul_step 9, H_POW1, GHASH_ACC, GHASH_ACC, GFPOLY, \
1071                         %xmm1, %xmm2, %xmm3
1072 
1073         // Undo the byte reflection of the GHASH accumulator.
1074         vpshufb         BSWAP_MASK, GHASH_ACC, GHASH_ACC
1075 
1076         // Do the last AES round and XOR the resulting keystream block with the
1077         // GHASH accumulator to produce the full computed authentication tag.
1078         //
1079         // Reduce latency by taking advantage of the property vaesenclast(key,
1080         // a) ^ b == vaesenclast(key ^ b, a).  I.e., XOR GHASH_ACC into the last
1081         // round key, instead of XOR'ing the final AES output with GHASH_ACC.
1082         //
1083         // enc_final then returns the computed auth tag, while dec_final
1084         // compares it with the transmitted one and returns a bool.  To compare
1085         // the tags, dec_final XORs them together and uses vptest to check
1086         // whether the result is all-zeroes.  This should be constant-time.
1087         // dec_final applies the vaesenclast optimization to this additional
1088         // value XOR'd too, using vpternlogd to XOR the last round key, GHASH
1089         // accumulator, and transmitted auth tag together in one instruction.
1090 .if \enc
1091         vpxor           (%rax), GHASH_ACC, %xmm1
1092         vaesenclast     %xmm1, %xmm0, GHASH_ACC
1093         vmovdqu         GHASH_ACC, (GHASH_ACC_PTR)
1094 .else
1095         vmovdqu         (TAG), %xmm1
1096         vpternlogd      $0x96, (%rax), GHASH_ACC, %xmm1
1097         vaesenclast     %xmm1, %xmm0, %xmm0
1098         xor             %eax, %eax
1099         vmovdqu8        %xmm0, %xmm0{%k1}{z}    // Truncate to TAGLEN bytes
1100         vptest          %xmm0, %xmm0
1101         sete            %al
1102 .endif
1103         // No need for vzeroupper here, since only used xmm registers were used.
1104         RET
1105 .endm
1106 
1107 _set_veclen 32
1108 SYM_FUNC_START(aes_gcm_precompute_vaes_avx10_256)
1109         _aes_gcm_precompute
1110 SYM_FUNC_END(aes_gcm_precompute_vaes_avx10_256)
1111 SYM_FUNC_START(aes_gcm_enc_update_vaes_avx10_256)
1112         _aes_gcm_update 1
1113 SYM_FUNC_END(aes_gcm_enc_update_vaes_avx10_256)
1114 SYM_FUNC_START(aes_gcm_dec_update_vaes_avx10_256)
1115         _aes_gcm_update 0
1116 SYM_FUNC_END(aes_gcm_dec_update_vaes_avx10_256)
1117 
1118 _set_veclen 64
1119 SYM_FUNC_START(aes_gcm_precompute_vaes_avx10_512)
1120         _aes_gcm_precompute
1121 SYM_FUNC_END(aes_gcm_precompute_vaes_avx10_512)
1122 SYM_FUNC_START(aes_gcm_enc_update_vaes_avx10_512)
1123         _aes_gcm_update 1
1124 SYM_FUNC_END(aes_gcm_enc_update_vaes_avx10_512)
1125 SYM_FUNC_START(aes_gcm_dec_update_vaes_avx10_512)
1126         _aes_gcm_update 0
1127 SYM_FUNC_END(aes_gcm_dec_update_vaes_avx10_512)
1128 
1129 // void aes_gcm_aad_update_vaes_avx10(const struct aes_gcm_key_avx10 *key,
1130 //                                    u8 ghash_acc[16],
1131 //                                    const u8 *aad, int aadlen);
1132 //
1133 // This function processes the AAD (Additional Authenticated Data) in GCM.
1134 // Using the key |key|, it updates the GHASH accumulator |ghash_acc| with the
1135 // data given by |aad| and |aadlen|.  |key->ghash_key_powers| must have been
1136 // initialized.  On the first call, |ghash_acc| must be all zeroes.  |aadlen|
1137 // must be a multiple of 16, except on the last call where it can be any length.
1138 // The caller must do any buffering needed to ensure this.
1139 //
1140 // AES-GCM is almost always used with small amounts of AAD, less than 32 bytes.
1141 // Therefore, for AAD processing we currently only provide this implementation
1142 // which uses 256-bit vectors (ymm registers) and only has a 1x-wide loop.  This
1143 // keeps the code size down, and it enables some micro-optimizations, e.g. using
1144 // VEX-coded instructions instead of EVEX-coded to save some instruction bytes.
1145 // To optimize for large amounts of AAD, we could implement a 4x-wide loop and
1146 // provide a version using 512-bit vectors, but that doesn't seem to be useful.
1147 SYM_FUNC_START(aes_gcm_aad_update_vaes_avx10)
1148 
1149         // Function arguments
1150         .set    KEY,            %rdi
1151         .set    GHASH_ACC_PTR,  %rsi
1152         .set    AAD,            %rdx
1153         .set    AADLEN,         %ecx
1154         .set    AADLEN64,       %rcx    // Zero-extend AADLEN before using!
1155 
1156         // Additional local variables.
1157         // %rax, %ymm0-%ymm3, and %k1 are used as temporary registers.
1158         .set    BSWAP_MASK,     %ymm4
1159         .set    GFPOLY,         %ymm5
1160         .set    GHASH_ACC,      %ymm6
1161         .set    GHASH_ACC_XMM,  %xmm6
1162         .set    H_POW1,         %ymm7
1163 
1164         // Load some constants.
1165         vbroadcasti128  .Lbswap_mask(%rip), BSWAP_MASK
1166         vbroadcasti128  .Lgfpoly(%rip), GFPOLY
1167 
1168         // Load the GHASH accumulator.
1169         vmovdqu         (GHASH_ACC_PTR), GHASH_ACC_XMM
1170 
1171         // Update GHASH with 32 bytes of AAD at a time.
1172         //
1173         // Pre-subtracting 32 from AADLEN saves an instruction from the loop and
1174         // also ensures that at least one write always occurs to AADLEN,
1175         // zero-extending it and allowing AADLEN64 to be used later.
1176         sub             $32, AADLEN
1177         jl              .Laad_loop_1x_done
1178         vmovdqu8        OFFSETOFEND_H_POWERS-32(KEY), H_POW1    // [H^2, H^1]
1179 .Laad_loop_1x:
1180         vmovdqu         (AAD), %ymm0
1181         vpshufb         BSWAP_MASK, %ymm0, %ymm0
1182         vpxor           %ymm0, GHASH_ACC, GHASH_ACC
1183         _ghash_mul      H_POW1, GHASH_ACC, GHASH_ACC, GFPOLY, \
1184                         %ymm0, %ymm1, %ymm2
1185         vextracti128    $1, GHASH_ACC, %xmm0
1186         vpxor           %xmm0, GHASH_ACC_XMM, GHASH_ACC_XMM
1187         add             $32, AAD
1188         sub             $32, AADLEN
1189         jge             .Laad_loop_1x
1190 .Laad_loop_1x_done:
1191         add             $32, AADLEN
1192         jz              .Laad_done
1193 
1194         // Update GHASH with the remaining 1 <= AADLEN < 32 bytes of AAD.
1195         mov             $-1, %eax
1196         bzhi            AADLEN, %eax, %eax
1197         kmovd           %eax, %k1
1198         vmovdqu8        (AAD), %ymm0{%k1}{z}
1199         neg             AADLEN64
1200         and             $~15, AADLEN64  // -round_up(AADLEN, 16)
1201         vmovdqu8        OFFSETOFEND_H_POWERS(KEY,AADLEN64), H_POW1
1202         vpshufb         BSWAP_MASK, %ymm0, %ymm0
1203         vpxor           %ymm0, GHASH_ACC, GHASH_ACC
1204         _ghash_mul      H_POW1, GHASH_ACC, GHASH_ACC, GFPOLY, \
1205                         %ymm0, %ymm1, %ymm2
1206         vextracti128    $1, GHASH_ACC, %xmm0
1207         vpxor           %xmm0, GHASH_ACC_XMM, GHASH_ACC_XMM
1208 
1209 .Laad_done:
1210         // Store the updated GHASH accumulator back to memory.
1211         vmovdqu         GHASH_ACC_XMM, (GHASH_ACC_PTR)
1212 
1213         vzeroupper      // This is needed after using ymm or zmm registers.
1214         RET
1215 SYM_FUNC_END(aes_gcm_aad_update_vaes_avx10)
1216 
1217 SYM_FUNC_START(aes_gcm_enc_final_vaes_avx10)
1218         _aes_gcm_final  1
1219 SYM_FUNC_END(aes_gcm_enc_final_vaes_avx10)
1220 SYM_FUNC_START(aes_gcm_dec_final_vaes_avx10)
1221         _aes_gcm_final  0
1222 SYM_FUNC_END(aes_gcm_dec_final_vaes_avx10)

~ [ source navigation ] ~ [ diff markup ] ~ [ identifier search ] ~

kernel.org | git.kernel.org | LWN.net | Project Home | SVN repository | Mail admin

Linux® is a registered trademark of Linus Torvalds in the United States and other countries.
TOMOYO® is a registered trademark of NTT DATA CORPORATION.

sflogo.php