~ [ source navigation ] ~ [ diff markup ] ~ [ identifier search ] ~

TOMOYO Linux Cross Reference
Linux/lib/crypto/mpi/mpih-div.c

Version: ~ [ linux-6.11.5 ] ~ [ linux-6.10.14 ] ~ [ linux-6.9.12 ] ~ [ linux-6.8.12 ] ~ [ linux-6.7.12 ] ~ [ linux-6.6.58 ] ~ [ linux-6.5.13 ] ~ [ linux-6.4.16 ] ~ [ linux-6.3.13 ] ~ [ linux-6.2.16 ] ~ [ linux-6.1.114 ] ~ [ linux-6.0.19 ] ~ [ linux-5.19.17 ] ~ [ linux-5.18.19 ] ~ [ linux-5.17.15 ] ~ [ linux-5.16.20 ] ~ [ linux-5.15.169 ] ~ [ linux-5.14.21 ] ~ [ linux-5.13.19 ] ~ [ linux-5.12.19 ] ~ [ linux-5.11.22 ] ~ [ linux-5.10.228 ] ~ [ linux-5.9.16 ] ~ [ linux-5.8.18 ] ~ [ linux-5.7.19 ] ~ [ linux-5.6.19 ] ~ [ linux-5.5.19 ] ~ [ linux-5.4.284 ] ~ [ linux-5.3.18 ] ~ [ linux-5.2.21 ] ~ [ linux-5.1.21 ] ~ [ linux-5.0.21 ] ~ [ linux-4.20.17 ] ~ [ linux-4.19.322 ] ~ [ linux-4.18.20 ] ~ [ linux-4.17.19 ] ~ [ linux-4.16.18 ] ~ [ linux-4.15.18 ] ~ [ linux-4.14.336 ] ~ [ linux-4.13.16 ] ~ [ linux-4.12.14 ] ~ [ linux-4.11.12 ] ~ [ linux-4.10.17 ] ~ [ linux-4.9.337 ] ~ [ linux-4.4.302 ] ~ [ linux-3.10.108 ] ~ [ linux-2.6.32.71 ] ~ [ linux-2.6.0 ] ~ [ linux-2.4.37.11 ] ~ [ unix-v6-master ] ~ [ ccs-tools-1.8.9 ] ~ [ policy-sample ] ~
Architecture: ~ [ i386 ] ~ [ alpha ] ~ [ m68k ] ~ [ mips ] ~ [ ppc ] ~ [ sparc ] ~ [ sparc64 ] ~

  1 // SPDX-License-Identifier: GPL-2.0-or-later
  2 /* mpihelp-div.c  -  MPI helper functions
  3  *      Copyright (C) 1994, 1996 Free Software Foundation, Inc.
  4  *      Copyright (C) 1998, 1999 Free Software Foundation, Inc.
  5  *
  6  * This file is part of GnuPG.
  7  *
  8  * Note: This code is heavily based on the GNU MP Library.
  9  *       Actually it's the same code with only minor changes in the
 10  *       way the data is stored; this is to support the abstraction
 11  *       of an optional secure memory allocation which may be used
 12  *       to avoid revealing of sensitive data due to paging etc.
 13  *       The GNU MP Library itself is published under the LGPL;
 14  *       however I decided to publish this code under the plain GPL.
 15  */
 16 
 17 #include "mpi-internal.h"
 18 #include "longlong.h"
 19 
 20 #ifndef UMUL_TIME
 21 #define UMUL_TIME 1
 22 #endif
 23 #ifndef UDIV_TIME
 24 #define UDIV_TIME UMUL_TIME
 25 #endif
 26 
 27 
 28 mpi_limb_t
 29 mpihelp_mod_1(mpi_ptr_t dividend_ptr, mpi_size_t dividend_size,
 30                         mpi_limb_t divisor_limb)
 31 {
 32         mpi_size_t i;
 33         mpi_limb_t n1, n0, r;
 34         mpi_limb_t dummy __maybe_unused;
 35 
 36         /* Botch: Should this be handled at all?  Rely on callers?      */
 37         if (!dividend_size)
 38                 return 0;
 39 
 40         /* If multiplication is much faster than division, and the
 41          * dividend is large, pre-invert the divisor, and use
 42          * only multiplications in the inner loop.
 43          *
 44          * This test should be read:
 45          *       Does it ever help to use udiv_qrnnd_preinv?
 46          *         && Does what we save compensate for the inversion overhead?
 47          */
 48         if (UDIV_TIME > (2 * UMUL_TIME + 6)
 49                         && (UDIV_TIME - (2 * UMUL_TIME + 6)) * dividend_size > UDIV_TIME) {
 50                 int normalization_steps;
 51 
 52                 normalization_steps = count_leading_zeros(divisor_limb);
 53                 if (normalization_steps) {
 54                         mpi_limb_t divisor_limb_inverted;
 55 
 56                         divisor_limb <<= normalization_steps;
 57 
 58                         /* Compute (2**2N - 2**N * DIVISOR_LIMB) / DIVISOR_LIMB.  The
 59                          * result is a (N+1)-bit approximation to 1/DIVISOR_LIMB, with the
 60                          * most significant bit (with weight 2**N) implicit.
 61                          *
 62                          * Special case for DIVISOR_LIMB == 100...000.
 63                          */
 64                         if (!(divisor_limb << 1))
 65                                 divisor_limb_inverted = ~(mpi_limb_t)0;
 66                         else
 67                                 udiv_qrnnd(divisor_limb_inverted, dummy,
 68                                                 -divisor_limb, 0, divisor_limb);
 69 
 70                         n1 = dividend_ptr[dividend_size - 1];
 71                         r = n1 >> (BITS_PER_MPI_LIMB - normalization_steps);
 72 
 73                         /* Possible optimization:
 74                          * if (r == 0
 75                          * && divisor_limb > ((n1 << normalization_steps)
 76                          *                     | (dividend_ptr[dividend_size - 2] >> ...)))
 77                          * ...one division less...
 78                          */
 79                         for (i = dividend_size - 2; i >= 0; i--) {
 80                                 n0 = dividend_ptr[i];
 81                                 UDIV_QRNND_PREINV(dummy, r, r,
 82                                                 ((n1 << normalization_steps)
 83                                                  | (n0 >> (BITS_PER_MPI_LIMB - normalization_steps))),
 84                                                 divisor_limb, divisor_limb_inverted);
 85                                 n1 = n0;
 86                         }
 87                         UDIV_QRNND_PREINV(dummy, r, r,
 88                                         n1 << normalization_steps,
 89                                         divisor_limb, divisor_limb_inverted);
 90                         return r >> normalization_steps;
 91                 } else {
 92                         mpi_limb_t divisor_limb_inverted;
 93 
 94                         /* Compute (2**2N - 2**N * DIVISOR_LIMB) / DIVISOR_LIMB.  The
 95                          * result is a (N+1)-bit approximation to 1/DIVISOR_LIMB, with the
 96                          * most significant bit (with weight 2**N) implicit.
 97                          *
 98                          * Special case for DIVISOR_LIMB == 100...000.
 99                          */
100                         if (!(divisor_limb << 1))
101                                 divisor_limb_inverted = ~(mpi_limb_t)0;
102                         else
103                                 udiv_qrnnd(divisor_limb_inverted, dummy,
104                                                 -divisor_limb, 0, divisor_limb);
105 
106                         i = dividend_size - 1;
107                         r = dividend_ptr[i];
108 
109                         if (r >= divisor_limb)
110                                 r = 0;
111                         else
112                                 i--;
113 
114                         for ( ; i >= 0; i--) {
115                                 n0 = dividend_ptr[i];
116                                 UDIV_QRNND_PREINV(dummy, r, r,
117                                                 n0, divisor_limb, divisor_limb_inverted);
118                         }
119                         return r;
120                 }
121         } else {
122                 if (UDIV_NEEDS_NORMALIZATION) {
123                         int normalization_steps;
124 
125                         normalization_steps = count_leading_zeros(divisor_limb);
126                         if (normalization_steps) {
127                                 divisor_limb <<= normalization_steps;
128 
129                                 n1 = dividend_ptr[dividend_size - 1];
130                                 r = n1 >> (BITS_PER_MPI_LIMB - normalization_steps);
131 
132                                 /* Possible optimization:
133                                  * if (r == 0
134                                  * && divisor_limb > ((n1 << normalization_steps)
135                                  *                 | (dividend_ptr[dividend_size - 2] >> ...)))
136                                  * ...one division less...
137                                  */
138                                 for (i = dividend_size - 2; i >= 0; i--) {
139                                         n0 = dividend_ptr[i];
140                                         udiv_qrnnd(dummy, r, r,
141                                                 ((n1 << normalization_steps)
142                                                  | (n0 >> (BITS_PER_MPI_LIMB - normalization_steps))),
143                                                 divisor_limb);
144                                         n1 = n0;
145                                 }
146                                 udiv_qrnnd(dummy, r, r,
147                                                 n1 << normalization_steps,
148                                                 divisor_limb);
149                                 return r >> normalization_steps;
150                         }
151                 }
152                 /* No normalization needed, either because udiv_qrnnd doesn't require
153                  * it, or because DIVISOR_LIMB is already normalized.
154                  */
155                 i = dividend_size - 1;
156                 r = dividend_ptr[i];
157 
158                 if (r >= divisor_limb)
159                         r = 0;
160                 else
161                         i--;
162 
163                 for (; i >= 0; i--) {
164                         n0 = dividend_ptr[i];
165                         udiv_qrnnd(dummy, r, r, n0, divisor_limb);
166                 }
167                 return r;
168         }
169 }
170 
171 /* Divide num (NP/NSIZE) by den (DP/DSIZE) and write
172  * the NSIZE-DSIZE least significant quotient limbs at QP
173  * and the DSIZE long remainder at NP.  If QEXTRA_LIMBS is
174  * non-zero, generate that many fraction bits and append them after the
175  * other quotient limbs.
176  * Return the most significant limb of the quotient, this is always 0 or 1.
177  *
178  * Preconditions:
179  * 0. NSIZE >= DSIZE.
180  * 1. The most significant bit of the divisor must be set.
181  * 2. QP must either not overlap with the input operands at all, or
182  *    QP + DSIZE >= NP must hold true.  (This means that it's
183  *    possible to put the quotient in the high part of NUM, right after the
184  *    remainder in NUM.
185  * 3. NSIZE >= DSIZE, even if QEXTRA_LIMBS is non-zero.
186  */
187 
188 mpi_limb_t
189 mpihelp_divrem(mpi_ptr_t qp, mpi_size_t qextra_limbs,
190                mpi_ptr_t np, mpi_size_t nsize, mpi_ptr_t dp, mpi_size_t dsize)
191 {
192         mpi_limb_t most_significant_q_limb = 0;
193 
194         switch (dsize) {
195         case 0:
196                 /* We are asked to divide by zero, so go ahead and do it!  (To make
197                    the compiler not remove this statement, return the value.)  */
198                 /*
199                  * existing clients of this function have been modified
200                  * not to call it with dsize == 0, so this should not happen
201                  */
202                 return 1 / dsize;
203 
204         case 1:
205                 {
206                         mpi_size_t i;
207                         mpi_limb_t n1;
208                         mpi_limb_t d;
209 
210                         d = dp[0];
211                         n1 = np[nsize - 1];
212 
213                         if (n1 >= d) {
214                                 n1 -= d;
215                                 most_significant_q_limb = 1;
216                         }
217 
218                         qp += qextra_limbs;
219                         for (i = nsize - 2; i >= 0; i--)
220                                 udiv_qrnnd(qp[i], n1, n1, np[i], d);
221                         qp -= qextra_limbs;
222 
223                         for (i = qextra_limbs - 1; i >= 0; i--)
224                                 udiv_qrnnd(qp[i], n1, n1, 0, d);
225 
226                         np[0] = n1;
227                 }
228                 break;
229 
230         case 2:
231                 {
232                         mpi_size_t i;
233                         mpi_limb_t n1, n0, n2;
234                         mpi_limb_t d1, d0;
235 
236                         np += nsize - 2;
237                         d1 = dp[1];
238                         d0 = dp[0];
239                         n1 = np[1];
240                         n0 = np[0];
241 
242                         if (n1 >= d1 && (n1 > d1 || n0 >= d0)) {
243                                 sub_ddmmss(n1, n0, n1, n0, d1, d0);
244                                 most_significant_q_limb = 1;
245                         }
246 
247                         for (i = qextra_limbs + nsize - 2 - 1; i >= 0; i--) {
248                                 mpi_limb_t q;
249                                 mpi_limb_t r;
250 
251                                 if (i >= qextra_limbs)
252                                         np--;
253                                 else
254                                         np[0] = 0;
255 
256                                 if (n1 == d1) {
257                                         /* Q should be either 111..111 or 111..110.  Need special
258                                          * treatment of this rare case as normal division would
259                                          * give overflow.  */
260                                         q = ~(mpi_limb_t) 0;
261 
262                                         r = n0 + d1;
263                                         if (r < d1) {   /* Carry in the addition? */
264                                                 add_ssaaaa(n1, n0, r - d0,
265                                                            np[0], 0, d0);
266                                                 qp[i] = q;
267                                                 continue;
268                                         }
269                                         n1 = d0 - (d0 != 0 ? 1 : 0);
270                                         n0 = -d0;
271                                 } else {
272                                         udiv_qrnnd(q, r, n1, n0, d1);
273                                         umul_ppmm(n1, n0, d0, q);
274                                 }
275 
276                                 n2 = np[0];
277 q_test:
278                                 if (n1 > r || (n1 == r && n0 > n2)) {
279                                         /* The estimated Q was too large.  */
280                                         q--;
281                                         sub_ddmmss(n1, n0, n1, n0, 0, d0);
282                                         r += d1;
283                                         if (r >= d1)    /* If not carry, test Q again.  */
284                                                 goto q_test;
285                                 }
286 
287                                 qp[i] = q;
288                                 sub_ddmmss(n1, n0, r, n2, n1, n0);
289                         }
290                         np[1] = n1;
291                         np[0] = n0;
292                 }
293                 break;
294 
295         default:
296                 {
297                         mpi_size_t i;
298                         mpi_limb_t dX, d1, n0;
299 
300                         np += nsize - dsize;
301                         dX = dp[dsize - 1];
302                         d1 = dp[dsize - 2];
303                         n0 = np[dsize - 1];
304 
305                         if (n0 >= dX) {
306                                 if (n0 > dX
307                                     || mpihelp_cmp(np, dp, dsize - 1) >= 0) {
308                                         mpihelp_sub_n(np, np, dp, dsize);
309                                         n0 = np[dsize - 1];
310                                         most_significant_q_limb = 1;
311                                 }
312                         }
313 
314                         for (i = qextra_limbs + nsize - dsize - 1; i >= 0; i--) {
315                                 mpi_limb_t q;
316                                 mpi_limb_t n1, n2;
317                                 mpi_limb_t cy_limb;
318 
319                                 if (i >= qextra_limbs) {
320                                         np--;
321                                         n2 = np[dsize];
322                                 } else {
323                                         n2 = np[dsize - 1];
324                                         MPN_COPY_DECR(np + 1, np, dsize - 1);
325                                         np[0] = 0;
326                                 }
327 
328                                 if (n0 == dX) {
329                                         /* This might over-estimate q, but it's probably not worth
330                                          * the extra code here to find out.  */
331                                         q = ~(mpi_limb_t) 0;
332                                 } else {
333                                         mpi_limb_t r;
334 
335                                         udiv_qrnnd(q, r, n0, np[dsize - 1], dX);
336                                         umul_ppmm(n1, n0, d1, q);
337 
338                                         while (n1 > r
339                                                || (n1 == r
340                                                    && n0 > np[dsize - 2])) {
341                                                 q--;
342                                                 r += dX;
343                                                 if (r < dX)     /* I.e. "carry in previous addition?" */
344                                                         break;
345                                                 n1 -= n0 < d1;
346                                                 n0 -= d1;
347                                         }
348                                 }
349 
350                                 /* Possible optimization: We already have (q * n0) and (1 * n1)
351                                  * after the calculation of q.  Taking advantage of that, we
352                                  * could make this loop make two iterations less.  */
353                                 cy_limb = mpihelp_submul_1(np, dp, dsize, q);
354 
355                                 if (n2 != cy_limb) {
356                                         mpihelp_add_n(np, np, dp, dsize);
357                                         q--;
358                                 }
359 
360                                 qp[i] = q;
361                                 n0 = np[dsize - 1];
362                         }
363                 }
364         }
365 
366         return most_significant_q_limb;
367 }
368 
369 /****************
370  * Divide (DIVIDEND_PTR,,DIVIDEND_SIZE) by DIVISOR_LIMB.
371  * Write DIVIDEND_SIZE limbs of quotient at QUOT_PTR.
372  * Return the single-limb remainder.
373  * There are no constraints on the value of the divisor.
374  *
375  * QUOT_PTR and DIVIDEND_PTR might point to the same limb.
376  */
377 
378 mpi_limb_t
379 mpihelp_divmod_1(mpi_ptr_t quot_ptr,
380                 mpi_ptr_t dividend_ptr, mpi_size_t dividend_size,
381                 mpi_limb_t divisor_limb)
382 {
383         mpi_size_t i;
384         mpi_limb_t n1, n0, r;
385         mpi_limb_t dummy __maybe_unused;
386 
387         if (!dividend_size)
388                 return 0;
389 
390         /* If multiplication is much faster than division, and the
391          * dividend is large, pre-invert the divisor, and use
392          * only multiplications in the inner loop.
393          *
394          * This test should be read:
395          * Does it ever help to use udiv_qrnnd_preinv?
396          * && Does what we save compensate for the inversion overhead?
397          */
398         if (UDIV_TIME > (2 * UMUL_TIME + 6)
399                         && (UDIV_TIME - (2 * UMUL_TIME + 6)) * dividend_size > UDIV_TIME) {
400                 int normalization_steps;
401 
402                 normalization_steps = count_leading_zeros(divisor_limb);
403                 if (normalization_steps) {
404                         mpi_limb_t divisor_limb_inverted;
405 
406                         divisor_limb <<= normalization_steps;
407 
408                         /* Compute (2**2N - 2**N * DIVISOR_LIMB) / DIVISOR_LIMB.  The
409                          * result is a (N+1)-bit approximation to 1/DIVISOR_LIMB, with the
410                          * most significant bit (with weight 2**N) implicit.
411                          */
412                         /* Special case for DIVISOR_LIMB == 100...000.  */
413                         if (!(divisor_limb << 1))
414                                 divisor_limb_inverted = ~(mpi_limb_t)0;
415                         else
416                                 udiv_qrnnd(divisor_limb_inverted, dummy,
417                                                 -divisor_limb, 0, divisor_limb);
418 
419                         n1 = dividend_ptr[dividend_size - 1];
420                         r = n1 >> (BITS_PER_MPI_LIMB - normalization_steps);
421 
422                         /* Possible optimization:
423                          * if (r == 0
424                          * && divisor_limb > ((n1 << normalization_steps)
425                          *                     | (dividend_ptr[dividend_size - 2] >> ...)))
426                          * ...one division less...
427                          */
428                         for (i = dividend_size - 2; i >= 0; i--) {
429                                 n0 = dividend_ptr[i];
430                                 UDIV_QRNND_PREINV(quot_ptr[i + 1], r, r,
431                                                 ((n1 << normalization_steps)
432                                                  | (n0 >> (BITS_PER_MPI_LIMB - normalization_steps))),
433                                                 divisor_limb, divisor_limb_inverted);
434                                 n1 = n0;
435                         }
436                         UDIV_QRNND_PREINV(quot_ptr[0], r, r,
437                                         n1 << normalization_steps,
438                                         divisor_limb, divisor_limb_inverted);
439                         return r >> normalization_steps;
440                 } else {
441                         mpi_limb_t divisor_limb_inverted;
442 
443                         /* Compute (2**2N - 2**N * DIVISOR_LIMB) / DIVISOR_LIMB.  The
444                          * result is a (N+1)-bit approximation to 1/DIVISOR_LIMB, with the
445                          * most significant bit (with weight 2**N) implicit.
446                          */
447                         /* Special case for DIVISOR_LIMB == 100...000.  */
448                         if (!(divisor_limb << 1))
449                                 divisor_limb_inverted = ~(mpi_limb_t) 0;
450                         else
451                                 udiv_qrnnd(divisor_limb_inverted, dummy,
452                                                 -divisor_limb, 0, divisor_limb);
453 
454                         i = dividend_size - 1;
455                         r = dividend_ptr[i];
456 
457                         if (r >= divisor_limb)
458                                 r = 0;
459                         else
460                                 quot_ptr[i--] = 0;
461 
462                         for ( ; i >= 0; i--) {
463                                 n0 = dividend_ptr[i];
464                                 UDIV_QRNND_PREINV(quot_ptr[i], r, r,
465                                                 n0, divisor_limb, divisor_limb_inverted);
466                         }
467                         return r;
468                 }
469         } else {
470                 if (UDIV_NEEDS_NORMALIZATION) {
471                         int normalization_steps;
472 
473                         normalization_steps = count_leading_zeros(divisor_limb);
474                         if (normalization_steps) {
475                                 divisor_limb <<= normalization_steps;
476 
477                                 n1 = dividend_ptr[dividend_size - 1];
478                                 r = n1 >> (BITS_PER_MPI_LIMB - normalization_steps);
479 
480                                 /* Possible optimization:
481                                  * if (r == 0
482                                  * && divisor_limb > ((n1 << normalization_steps)
483                                  *                 | (dividend_ptr[dividend_size - 2] >> ...)))
484                                  * ...one division less...
485                                  */
486                                 for (i = dividend_size - 2; i >= 0; i--) {
487                                         n0 = dividend_ptr[i];
488                                         udiv_qrnnd(quot_ptr[i + 1], r, r,
489                                                 ((n1 << normalization_steps)
490                                                  | (n0 >> (BITS_PER_MPI_LIMB - normalization_steps))),
491                                                 divisor_limb);
492                                         n1 = n0;
493                                 }
494                                 udiv_qrnnd(quot_ptr[0], r, r,
495                                                 n1 << normalization_steps,
496                                                 divisor_limb);
497                                 return r >> normalization_steps;
498                         }
499                 }
500                 /* No normalization needed, either because udiv_qrnnd doesn't require
501                  * it, or because DIVISOR_LIMB is already normalized.
502                  */
503                 i = dividend_size - 1;
504                 r = dividend_ptr[i];
505 
506                 if (r >= divisor_limb)
507                         r = 0;
508                 else
509                         quot_ptr[i--] = 0;
510 
511                 for (; i >= 0; i--) {
512                         n0 = dividend_ptr[i];
513                         udiv_qrnnd(quot_ptr[i], r, r, n0, divisor_limb);
514                 }
515                 return r;
516         }
517 }
518 

~ [ source navigation ] ~ [ diff markup ] ~ [ identifier search ] ~

kernel.org | git.kernel.org | LWN.net | Project Home | SVN repository | Mail admin

Linux® is a registered trademark of Linus Torvalds in the United States and other countries.
TOMOYO® is a registered trademark of NTT DATA CORPORATION.

sflogo.php