~ [ source navigation ] ~ [ diff markup ] ~ [ identifier search ] ~

TOMOYO Linux Cross Reference
Linux/arch/sh/kernel/cpu/sh4/softfloat.c

Version: ~ [ linux-6.11.5 ] ~ [ linux-6.10.14 ] ~ [ linux-6.9.12 ] ~ [ linux-6.8.12 ] ~ [ linux-6.7.12 ] ~ [ linux-6.6.58 ] ~ [ linux-6.5.13 ] ~ [ linux-6.4.16 ] ~ [ linux-6.3.13 ] ~ [ linux-6.2.16 ] ~ [ linux-6.1.114 ] ~ [ linux-6.0.19 ] ~ [ linux-5.19.17 ] ~ [ linux-5.18.19 ] ~ [ linux-5.17.15 ] ~ [ linux-5.16.20 ] ~ [ linux-5.15.169 ] ~ [ linux-5.14.21 ] ~ [ linux-5.13.19 ] ~ [ linux-5.12.19 ] ~ [ linux-5.11.22 ] ~ [ linux-5.10.228 ] ~ [ linux-5.9.16 ] ~ [ linux-5.8.18 ] ~ [ linux-5.7.19 ] ~ [ linux-5.6.19 ] ~ [ linux-5.5.19 ] ~ [ linux-5.4.284 ] ~ [ linux-5.3.18 ] ~ [ linux-5.2.21 ] ~ [ linux-5.1.21 ] ~ [ linux-5.0.21 ] ~ [ linux-4.20.17 ] ~ [ linux-4.19.322 ] ~ [ linux-4.18.20 ] ~ [ linux-4.17.19 ] ~ [ linux-4.16.18 ] ~ [ linux-4.15.18 ] ~ [ linux-4.14.336 ] ~ [ linux-4.13.16 ] ~ [ linux-4.12.14 ] ~ [ linux-4.11.12 ] ~ [ linux-4.10.17 ] ~ [ linux-4.9.337 ] ~ [ linux-4.4.302 ] ~ [ linux-3.10.108 ] ~ [ linux-2.6.32.71 ] ~ [ linux-2.6.0 ] ~ [ linux-2.4.37.11 ] ~ [ unix-v6-master ] ~ [ ccs-tools-1.8.9 ] ~ [ policy-sample ] ~
Architecture: ~ [ i386 ] ~ [ alpha ] ~ [ m68k ] ~ [ mips ] ~ [ ppc ] ~ [ sparc ] ~ [ sparc64 ] ~

  1 /*
  2  * Floating point emulation support for subnormalised numbers on SH4
  3  * architecture This file is derived from the SoftFloat IEC/IEEE
  4  * Floating-point Arithmetic Package, Release 2 the original license of
  5  * which is reproduced below.
  6  *
  7  * ========================================================================
  8  *
  9  * This C source file is part of the SoftFloat IEC/IEEE Floating-point
 10  * Arithmetic Package, Release 2.
 11  *
 12  * Written by John R. Hauser.  This work was made possible in part by the
 13  * International Computer Science Institute, located at Suite 600, 1947 Center
 14  * Street, Berkeley, California 94704.  Funding was partially provided by the
 15  * National Science Foundation under grant MIP-9311980.  The original version
 16  * of this code was written as part of a project to build a fixed-point vector
 17  * processor in collaboration with the University of California at Berkeley,
 18  * overseen by Profs. Nelson Morgan and John Wawrzynek.  More information
 19  * is available through the web page `http://HTTP.CS.Berkeley.EDU/~jhauser/
 20  * arithmetic/softfloat.html'.
 21  *
 22  * THIS SOFTWARE IS DISTRIBUTED AS IS, FOR FREE.  Although reasonable effort
 23  * has been made to avoid it, THIS SOFTWARE MAY CONTAIN FAULTS THAT WILL AT
 24  * TIMES RESULT IN INCORRECT BEHAVIOR.  USE OF THIS SOFTWARE IS RESTRICTED TO
 25  * PERSONS AND ORGANIZATIONS WHO CAN AND WILL TAKE FULL RESPONSIBILITY FOR ANY
 26  * AND ALL LOSSES, COSTS, OR OTHER PROBLEMS ARISING FROM ITS USE.
 27  *
 28  * Derivative works are acceptable, even for commercial purposes, so long as
 29  * (1) they include prominent notice that the work is derivative, and (2) they
 30  * include prominent notice akin to these three paragraphs for those parts of
 31  * this code that are retained.
 32  *
 33  * ========================================================================
 34  *
 35  * SH4 modifications by Ismail Dhaoui <ismail.dhaoui@st.com>
 36  * and Kamel Khelifi <kamel.khelifi@st.com>
 37  */
 38 #include <linux/kernel.h>
 39 #include <cpu/fpu.h>
 40 #include <asm/div64.h>
 41 
 42 #define LIT64( a ) a##LL
 43 
 44 typedef char flag;
 45 typedef unsigned char uint8;
 46 typedef signed char int8;
 47 typedef int uint16;
 48 typedef int int16;
 49 typedef unsigned int uint32;
 50 typedef signed int int32;
 51 
 52 typedef unsigned long long int bits64;
 53 typedef signed long long int sbits64;
 54 
 55 typedef unsigned char bits8;
 56 typedef signed char sbits8;
 57 typedef unsigned short int bits16;
 58 typedef signed short int sbits16;
 59 typedef unsigned int bits32;
 60 typedef signed int sbits32;
 61 
 62 typedef unsigned long long int uint64;
 63 typedef signed long long int int64;
 64 
 65 typedef unsigned long int float32;
 66 typedef unsigned long long float64;
 67 
 68 extern void float_raise(unsigned int flags);    /* in fpu.c */
 69 extern int float_rounding_mode(void);   /* in fpu.c */
 70 
 71 bits64 extractFloat64Frac(float64 a);
 72 flag extractFloat64Sign(float64 a);
 73 int16 extractFloat64Exp(float64 a);
 74 int16 extractFloat32Exp(float32 a);
 75 flag extractFloat32Sign(float32 a);
 76 bits32 extractFloat32Frac(float32 a);
 77 float64 packFloat64(flag zSign, int16 zExp, bits64 zSig);
 78 void shift64RightJamming(bits64 a, int16 count, bits64 * zPtr);
 79 float32 packFloat32(flag zSign, int16 zExp, bits32 zSig);
 80 void shift32RightJamming(bits32 a, int16 count, bits32 * zPtr);
 81 float64 float64_sub(float64 a, float64 b);
 82 float32 float32_sub(float32 a, float32 b);
 83 float32 float32_add(float32 a, float32 b);
 84 float64 float64_add(float64 a, float64 b);
 85 float64 float64_div(float64 a, float64 b);
 86 float32 float32_div(float32 a, float32 b);
 87 float32 float32_mul(float32 a, float32 b);
 88 float64 float64_mul(float64 a, float64 b);
 89 float32 float64_to_float32(float64 a);
 90 void add128(bits64 a0, bits64 a1, bits64 b0, bits64 b1, bits64 * z0Ptr,
 91                    bits64 * z1Ptr);
 92 void sub128(bits64 a0, bits64 a1, bits64 b0, bits64 b1, bits64 * z0Ptr,
 93                    bits64 * z1Ptr);
 94 void mul64To128(bits64 a, bits64 b, bits64 * z0Ptr, bits64 * z1Ptr);
 95 
 96 static int8 countLeadingZeros32(bits32 a);
 97 static int8 countLeadingZeros64(bits64 a);
 98 static float64 normalizeRoundAndPackFloat64(flag zSign, int16 zExp,
 99                                             bits64 zSig);
100 static float64 subFloat64Sigs(float64 a, float64 b, flag zSign);
101 static float64 addFloat64Sigs(float64 a, float64 b, flag zSign);
102 static float32 roundAndPackFloat32(flag zSign, int16 zExp, bits32 zSig);
103 static float32 normalizeRoundAndPackFloat32(flag zSign, int16 zExp,
104                                             bits32 zSig);
105 static float64 roundAndPackFloat64(flag zSign, int16 zExp, bits64 zSig);
106 static float32 subFloat32Sigs(float32 a, float32 b, flag zSign);
107 static float32 addFloat32Sigs(float32 a, float32 b, flag zSign);
108 static void normalizeFloat64Subnormal(bits64 aSig, int16 * zExpPtr,
109                                       bits64 * zSigPtr);
110 static bits64 estimateDiv128To64(bits64 a0, bits64 a1, bits64 b);
111 static void normalizeFloat32Subnormal(bits32 aSig, int16 * zExpPtr,
112                                       bits32 * zSigPtr);
113 
114 bits64 extractFloat64Frac(float64 a)
115 {
116         return a & LIT64(0x000FFFFFFFFFFFFF);
117 }
118 
119 flag extractFloat64Sign(float64 a)
120 {
121         return a >> 63;
122 }
123 
124 int16 extractFloat64Exp(float64 a)
125 {
126         return (a >> 52) & 0x7FF;
127 }
128 
129 int16 extractFloat32Exp(float32 a)
130 {
131         return (a >> 23) & 0xFF;
132 }
133 
134 flag extractFloat32Sign(float32 a)
135 {
136         return a >> 31;
137 }
138 
139 bits32 extractFloat32Frac(float32 a)
140 {
141         return a & 0x007FFFFF;
142 }
143 
144 float64 packFloat64(flag zSign, int16 zExp, bits64 zSig)
145 {
146         return (((bits64) zSign) << 63) + (((bits64) zExp) << 52) + zSig;
147 }
148 
149 void shift64RightJamming(bits64 a, int16 count, bits64 * zPtr)
150 {
151         bits64 z;
152 
153         if (count == 0) {
154                 z = a;
155         } else if (count < 64) {
156                 z = (a >> count) | ((a << ((-count) & 63)) != 0);
157         } else {
158                 z = (a != 0);
159         }
160         *zPtr = z;
161 }
162 
163 static int8 countLeadingZeros32(bits32 a)
164 {
165         static const int8 countLeadingZerosHigh[] = {
166                 8, 7, 6, 6, 5, 5, 5, 5, 4, 4, 4, 4, 4, 4, 4, 4,
167                 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3,
168                 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2,
169                 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2,
170                 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1,
171                 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1,
172                 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1,
173                 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1,
174                 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
175                 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
176                 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
177                 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
178                 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
179                 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
180                 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
181                 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0
182         };
183         int8 shiftCount;
184 
185         shiftCount = 0;
186         if (a < 0x10000) {
187                 shiftCount += 16;
188                 a <<= 16;
189         }
190         if (a < 0x1000000) {
191                 shiftCount += 8;
192                 a <<= 8;
193         }
194         shiftCount += countLeadingZerosHigh[a >> 24];
195         return shiftCount;
196 
197 }
198 
199 static int8 countLeadingZeros64(bits64 a)
200 {
201         int8 shiftCount;
202 
203         shiftCount = 0;
204         if (a < ((bits64) 1) << 32) {
205                 shiftCount += 32;
206         } else {
207                 a >>= 32;
208         }
209         shiftCount += countLeadingZeros32(a);
210         return shiftCount;
211 
212 }
213 
214 static float64 normalizeRoundAndPackFloat64(flag zSign, int16 zExp, bits64 zSig)
215 {
216         int8 shiftCount;
217 
218         shiftCount = countLeadingZeros64(zSig) - 1;
219         return roundAndPackFloat64(zSign, zExp - shiftCount,
220                                    zSig << shiftCount);
221 
222 }
223 
224 static float64 subFloat64Sigs(float64 a, float64 b, flag zSign)
225 {
226         int16 aExp, bExp, zExp;
227         bits64 aSig, bSig, zSig;
228         int16 expDiff;
229 
230         aSig = extractFloat64Frac(a);
231         aExp = extractFloat64Exp(a);
232         bSig = extractFloat64Frac(b);
233         bExp = extractFloat64Exp(b);
234         expDiff = aExp - bExp;
235         aSig <<= 10;
236         bSig <<= 10;
237         if (0 < expDiff)
238                 goto aExpBigger;
239         if (expDiff < 0)
240                 goto bExpBigger;
241         if (aExp == 0) {
242                 aExp = 1;
243                 bExp = 1;
244         }
245         if (bSig < aSig)
246                 goto aBigger;
247         if (aSig < bSig)
248                 goto bBigger;
249         return packFloat64(float_rounding_mode() == FPSCR_RM_ZERO, 0, 0);
250       bExpBigger:
251         if (bExp == 0x7FF) {
252                 return packFloat64(zSign ^ 1, 0x7FF, 0);
253         }
254         if (aExp == 0) {
255                 ++expDiff;
256         } else {
257                 aSig |= LIT64(0x4000000000000000);
258         }
259         shift64RightJamming(aSig, -expDiff, &aSig);
260         bSig |= LIT64(0x4000000000000000);
261       bBigger:
262         zSig = bSig - aSig;
263         zExp = bExp;
264         zSign ^= 1;
265         goto normalizeRoundAndPack;
266       aExpBigger:
267         if (aExp == 0x7FF) {
268                 return a;
269         }
270         if (bExp == 0) {
271                 --expDiff;
272         } else {
273                 bSig |= LIT64(0x4000000000000000);
274         }
275         shift64RightJamming(bSig, expDiff, &bSig);
276         aSig |= LIT64(0x4000000000000000);
277       aBigger:
278         zSig = aSig - bSig;
279         zExp = aExp;
280       normalizeRoundAndPack:
281         --zExp;
282         return normalizeRoundAndPackFloat64(zSign, zExp, zSig);
283 
284 }
285 static float64 addFloat64Sigs(float64 a, float64 b, flag zSign)
286 {
287         int16 aExp, bExp, zExp;
288         bits64 aSig, bSig, zSig;
289         int16 expDiff;
290 
291         aSig = extractFloat64Frac(a);
292         aExp = extractFloat64Exp(a);
293         bSig = extractFloat64Frac(b);
294         bExp = extractFloat64Exp(b);
295         expDiff = aExp - bExp;
296         aSig <<= 9;
297         bSig <<= 9;
298         if (0 < expDiff) {
299                 if (aExp == 0x7FF) {
300                         return a;
301                 }
302                 if (bExp == 0) {
303                         --expDiff;
304                 } else {
305                         bSig |= LIT64(0x2000000000000000);
306                 }
307                 shift64RightJamming(bSig, expDiff, &bSig);
308                 zExp = aExp;
309         } else if (expDiff < 0) {
310                 if (bExp == 0x7FF) {
311                         return packFloat64(zSign, 0x7FF, 0);
312                 }
313                 if (aExp == 0) {
314                         ++expDiff;
315                 } else {
316                         aSig |= LIT64(0x2000000000000000);
317                 }
318                 shift64RightJamming(aSig, -expDiff, &aSig);
319                 zExp = bExp;
320         } else {
321                 if (aExp == 0x7FF) {
322                         return a;
323                 }
324                 if (aExp == 0)
325                         return packFloat64(zSign, 0, (aSig + bSig) >> 9);
326                 zSig = LIT64(0x4000000000000000) + aSig + bSig;
327                 zExp = aExp;
328                 goto roundAndPack;
329         }
330         aSig |= LIT64(0x2000000000000000);
331         zSig = (aSig + bSig) << 1;
332         --zExp;
333         if ((sbits64) zSig < 0) {
334                 zSig = aSig + bSig;
335                 ++zExp;
336         }
337       roundAndPack:
338         return roundAndPackFloat64(zSign, zExp, zSig);
339 
340 }
341 
342 float32 packFloat32(flag zSign, int16 zExp, bits32 zSig)
343 {
344         return (((bits32) zSign) << 31) + (((bits32) zExp) << 23) + zSig;
345 }
346 
347 void shift32RightJamming(bits32 a, int16 count, bits32 * zPtr)
348 {
349         bits32 z;
350         if (count == 0) {
351                 z = a;
352         } else if (count < 32) {
353                 z = (a >> count) | ((a << ((-count) & 31)) != 0);
354         } else {
355                 z = (a != 0);
356         }
357         *zPtr = z;
358 }
359 
360 static float32 roundAndPackFloat32(flag zSign, int16 zExp, bits32 zSig)
361 {
362         flag roundNearestEven;
363         int8 roundIncrement, roundBits;
364         flag isTiny;
365 
366         /* SH4 has only 2 rounding modes - round to nearest and round to zero */
367         roundNearestEven = (float_rounding_mode() == FPSCR_RM_NEAREST);
368         roundIncrement = 0x40;
369         if (!roundNearestEven) {
370                 roundIncrement = 0;
371         }
372         roundBits = zSig & 0x7F;
373         if (0xFD <= (bits16) zExp) {
374                 if ((0xFD < zExp)
375                     || ((zExp == 0xFD)
376                         && ((sbits32) (zSig + roundIncrement) < 0))
377                     ) {
378                         float_raise(FPSCR_CAUSE_OVERFLOW | FPSCR_CAUSE_INEXACT);
379                         return packFloat32(zSign, 0xFF,
380                                            0) - (roundIncrement == 0);
381                 }
382                 if (zExp < 0) {
383                         isTiny = (zExp < -1)
384                             || (zSig + roundIncrement < 0x80000000);
385                         shift32RightJamming(zSig, -zExp, &zSig);
386                         zExp = 0;
387                         roundBits = zSig & 0x7F;
388                         if (isTiny && roundBits)
389                                 float_raise(FPSCR_CAUSE_UNDERFLOW);
390                 }
391         }
392         if (roundBits)
393                 float_raise(FPSCR_CAUSE_INEXACT);
394         zSig = (zSig + roundIncrement) >> 7;
395         zSig &= ~(((roundBits ^ 0x40) == 0) & roundNearestEven);
396         if (zSig == 0)
397                 zExp = 0;
398         return packFloat32(zSign, zExp, zSig);
399 
400 }
401 
402 static float32 normalizeRoundAndPackFloat32(flag zSign, int16 zExp, bits32 zSig)
403 {
404         int8 shiftCount;
405 
406         shiftCount = countLeadingZeros32(zSig) - 1;
407         return roundAndPackFloat32(zSign, zExp - shiftCount,
408                                    zSig << shiftCount);
409 }
410 
411 static float64 roundAndPackFloat64(flag zSign, int16 zExp, bits64 zSig)
412 {
413         flag roundNearestEven;
414         int16 roundIncrement, roundBits;
415         flag isTiny;
416 
417         /* SH4 has only 2 rounding modes - round to nearest and round to zero */
418         roundNearestEven = (float_rounding_mode() == FPSCR_RM_NEAREST);
419         roundIncrement = 0x200;
420         if (!roundNearestEven) {
421                 roundIncrement = 0;
422         }
423         roundBits = zSig & 0x3FF;
424         if (0x7FD <= (bits16) zExp) {
425                 if ((0x7FD < zExp)
426                     || ((zExp == 0x7FD)
427                         && ((sbits64) (zSig + roundIncrement) < 0))
428                     ) {
429                         float_raise(FPSCR_CAUSE_OVERFLOW | FPSCR_CAUSE_INEXACT);
430                         return packFloat64(zSign, 0x7FF,
431                                            0) - (roundIncrement == 0);
432                 }
433                 if (zExp < 0) {
434                         isTiny = (zExp < -1)
435                             || (zSig + roundIncrement <
436                                 LIT64(0x8000000000000000));
437                         shift64RightJamming(zSig, -zExp, &zSig);
438                         zExp = 0;
439                         roundBits = zSig & 0x3FF;
440                         if (isTiny && roundBits)
441                                 float_raise(FPSCR_CAUSE_UNDERFLOW);
442                 }
443         }
444         if (roundBits)
445                 float_raise(FPSCR_CAUSE_INEXACT);
446         zSig = (zSig + roundIncrement) >> 10;
447         zSig &= ~(((roundBits ^ 0x200) == 0) & roundNearestEven);
448         if (zSig == 0)
449                 zExp = 0;
450         return packFloat64(zSign, zExp, zSig);
451 
452 }
453 
454 static float32 subFloat32Sigs(float32 a, float32 b, flag zSign)
455 {
456         int16 aExp, bExp, zExp;
457         bits32 aSig, bSig, zSig;
458         int16 expDiff;
459 
460         aSig = extractFloat32Frac(a);
461         aExp = extractFloat32Exp(a);
462         bSig = extractFloat32Frac(b);
463         bExp = extractFloat32Exp(b);
464         expDiff = aExp - bExp;
465         aSig <<= 7;
466         bSig <<= 7;
467         if (0 < expDiff)
468                 goto aExpBigger;
469         if (expDiff < 0)
470                 goto bExpBigger;
471         if (aExp == 0) {
472                 aExp = 1;
473                 bExp = 1;
474         }
475         if (bSig < aSig)
476                 goto aBigger;
477         if (aSig < bSig)
478                 goto bBigger;
479         return packFloat32(float_rounding_mode() == FPSCR_RM_ZERO, 0, 0);
480       bExpBigger:
481         if (bExp == 0xFF) {
482                 return packFloat32(zSign ^ 1, 0xFF, 0);
483         }
484         if (aExp == 0) {
485                 ++expDiff;
486         } else {
487                 aSig |= 0x40000000;
488         }
489         shift32RightJamming(aSig, -expDiff, &aSig);
490         bSig |= 0x40000000;
491       bBigger:
492         zSig = bSig - aSig;
493         zExp = bExp;
494         zSign ^= 1;
495         goto normalizeRoundAndPack;
496       aExpBigger:
497         if (aExp == 0xFF) {
498                 return a;
499         }
500         if (bExp == 0) {
501                 --expDiff;
502         } else {
503                 bSig |= 0x40000000;
504         }
505         shift32RightJamming(bSig, expDiff, &bSig);
506         aSig |= 0x40000000;
507       aBigger:
508         zSig = aSig - bSig;
509         zExp = aExp;
510       normalizeRoundAndPack:
511         --zExp;
512         return normalizeRoundAndPackFloat32(zSign, zExp, zSig);
513 
514 }
515 
516 static float32 addFloat32Sigs(float32 a, float32 b, flag zSign)
517 {
518         int16 aExp, bExp, zExp;
519         bits32 aSig, bSig, zSig;
520         int16 expDiff;
521 
522         aSig = extractFloat32Frac(a);
523         aExp = extractFloat32Exp(a);
524         bSig = extractFloat32Frac(b);
525         bExp = extractFloat32Exp(b);
526         expDiff = aExp - bExp;
527         aSig <<= 6;
528         bSig <<= 6;
529         if (0 < expDiff) {
530                 if (aExp == 0xFF) {
531                         return a;
532                 }
533                 if (bExp == 0) {
534                         --expDiff;
535                 } else {
536                         bSig |= 0x20000000;
537                 }
538                 shift32RightJamming(bSig, expDiff, &bSig);
539                 zExp = aExp;
540         } else if (expDiff < 0) {
541                 if (bExp == 0xFF) {
542                         return packFloat32(zSign, 0xFF, 0);
543                 }
544                 if (aExp == 0) {
545                         ++expDiff;
546                 } else {
547                         aSig |= 0x20000000;
548                 }
549                 shift32RightJamming(aSig, -expDiff, &aSig);
550                 zExp = bExp;
551         } else {
552                 if (aExp == 0xFF) {
553                         return a;
554                 }
555                 if (aExp == 0)
556                         return packFloat32(zSign, 0, (aSig + bSig) >> 6);
557                 zSig = 0x40000000 + aSig + bSig;
558                 zExp = aExp;
559                 goto roundAndPack;
560         }
561         aSig |= 0x20000000;
562         zSig = (aSig + bSig) << 1;
563         --zExp;
564         if ((sbits32) zSig < 0) {
565                 zSig = aSig + bSig;
566                 ++zExp;
567         }
568       roundAndPack:
569         return roundAndPackFloat32(zSign, zExp, zSig);
570 
571 }
572 
573 float64 float64_sub(float64 a, float64 b)
574 {
575         flag aSign, bSign;
576 
577         aSign = extractFloat64Sign(a);
578         bSign = extractFloat64Sign(b);
579         if (aSign == bSign) {
580                 return subFloat64Sigs(a, b, aSign);
581         } else {
582                 return addFloat64Sigs(a, b, aSign);
583         }
584 
585 }
586 
587 float32 float32_sub(float32 a, float32 b)
588 {
589         flag aSign, bSign;
590 
591         aSign = extractFloat32Sign(a);
592         bSign = extractFloat32Sign(b);
593         if (aSign == bSign) {
594                 return subFloat32Sigs(a, b, aSign);
595         } else {
596                 return addFloat32Sigs(a, b, aSign);
597         }
598 
599 }
600 
601 float32 float32_add(float32 a, float32 b)
602 {
603         flag aSign, bSign;
604 
605         aSign = extractFloat32Sign(a);
606         bSign = extractFloat32Sign(b);
607         if (aSign == bSign) {
608                 return addFloat32Sigs(a, b, aSign);
609         } else {
610                 return subFloat32Sigs(a, b, aSign);
611         }
612 
613 }
614 
615 float64 float64_add(float64 a, float64 b)
616 {
617         flag aSign, bSign;
618 
619         aSign = extractFloat64Sign(a);
620         bSign = extractFloat64Sign(b);
621         if (aSign == bSign) {
622                 return addFloat64Sigs(a, b, aSign);
623         } else {
624                 return subFloat64Sigs(a, b, aSign);
625         }
626 }
627 
628 static void
629 normalizeFloat64Subnormal(bits64 aSig, int16 * zExpPtr, bits64 * zSigPtr)
630 {
631         int8 shiftCount;
632 
633         shiftCount = countLeadingZeros64(aSig) - 11;
634         *zSigPtr = aSig << shiftCount;
635         *zExpPtr = 1 - shiftCount;
636 }
637 
638 void add128(bits64 a0, bits64 a1, bits64 b0, bits64 b1, bits64 * z0Ptr,
639                    bits64 * z1Ptr)
640 {
641         bits64 z1;
642 
643         z1 = a1 + b1;
644         *z1Ptr = z1;
645         *z0Ptr = a0 + b0 + (z1 < a1);
646 }
647 
648 void
649 sub128(bits64 a0, bits64 a1, bits64 b0, bits64 b1, bits64 * z0Ptr,
650        bits64 * z1Ptr)
651 {
652         *z1Ptr = a1 - b1;
653         *z0Ptr = a0 - b0 - (a1 < b1);
654 }
655 
656 static bits64 estimateDiv128To64(bits64 a0, bits64 a1, bits64 b)
657 {
658         bits64 b0, b1;
659         bits64 rem0, rem1, term0, term1;
660         bits64 z, tmp;
661         if (b <= a0)
662                 return LIT64(0xFFFFFFFFFFFFFFFF);
663         b0 = b >> 32;
664         tmp = a0;
665         do_div(tmp, b0);
666 
667         z = (b0 << 32 <= a0) ? LIT64(0xFFFFFFFF00000000) : tmp << 32;
668         mul64To128(b, z, &term0, &term1);
669         sub128(a0, a1, term0, term1, &rem0, &rem1);
670         while (((sbits64) rem0) < 0) {
671                 z -= LIT64(0x100000000);
672                 b1 = b << 32;
673                 add128(rem0, rem1, b0, b1, &rem0, &rem1);
674         }
675         rem0 = (rem0 << 32) | (rem1 >> 32);
676         tmp = rem0;
677         do_div(tmp, b0);
678         z |= (b0 << 32 <= rem0) ? 0xFFFFFFFF : tmp;
679         return z;
680 }
681 
682 void mul64To128(bits64 a, bits64 b, bits64 * z0Ptr, bits64 * z1Ptr)
683 {
684         bits32 aHigh, aLow, bHigh, bLow;
685         bits64 z0, zMiddleA, zMiddleB, z1;
686 
687         aLow = a;
688         aHigh = a >> 32;
689         bLow = b;
690         bHigh = b >> 32;
691         z1 = ((bits64) aLow) * bLow;
692         zMiddleA = ((bits64) aLow) * bHigh;
693         zMiddleB = ((bits64) aHigh) * bLow;
694         z0 = ((bits64) aHigh) * bHigh;
695         zMiddleA += zMiddleB;
696         z0 += (((bits64) (zMiddleA < zMiddleB)) << 32) + (zMiddleA >> 32);
697         zMiddleA <<= 32;
698         z1 += zMiddleA;
699         z0 += (z1 < zMiddleA);
700         *z1Ptr = z1;
701         *z0Ptr = z0;
702 
703 }
704 
705 static void normalizeFloat32Subnormal(bits32 aSig, int16 * zExpPtr,
706                                       bits32 * zSigPtr)
707 {
708         int8 shiftCount;
709 
710         shiftCount = countLeadingZeros32(aSig) - 8;
711         *zSigPtr = aSig << shiftCount;
712         *zExpPtr = 1 - shiftCount;
713 
714 }
715 
716 float64 float64_div(float64 a, float64 b)
717 {
718         flag aSign, bSign, zSign;
719         int16 aExp, bExp, zExp;
720         bits64 aSig, bSig, zSig;
721         bits64 rem0, rem1;
722         bits64 term0, term1;
723 
724         aSig = extractFloat64Frac(a);
725         aExp = extractFloat64Exp(a);
726         aSign = extractFloat64Sign(a);
727         bSig = extractFloat64Frac(b);
728         bExp = extractFloat64Exp(b);
729         bSign = extractFloat64Sign(b);
730         zSign = aSign ^ bSign;
731         if (aExp == 0x7FF) {
732                 if (bExp == 0x7FF) {
733                 }
734                 return packFloat64(zSign, 0x7FF, 0);
735         }
736         if (bExp == 0x7FF) {
737                 return packFloat64(zSign, 0, 0);
738         }
739         if (bExp == 0) {
740                 if (bSig == 0) {
741                         if ((aExp | aSig) == 0) {
742                                 float_raise(FPSCR_CAUSE_INVALID);
743                         }
744                         return packFloat64(zSign, 0x7FF, 0);
745                 }
746                 normalizeFloat64Subnormal(bSig, &bExp, &bSig);
747         }
748         if (aExp == 0) {
749                 if (aSig == 0)
750                         return packFloat64(zSign, 0, 0);
751                 normalizeFloat64Subnormal(aSig, &aExp, &aSig);
752         }
753         zExp = aExp - bExp + 0x3FD;
754         aSig = (aSig | LIT64(0x0010000000000000)) << 10;
755         bSig = (bSig | LIT64(0x0010000000000000)) << 11;
756         if (bSig <= (aSig + aSig)) {
757                 aSig >>= 1;
758                 ++zExp;
759         }
760         zSig = estimateDiv128To64(aSig, 0, bSig);
761         if ((zSig & 0x1FF) <= 2) {
762                 mul64To128(bSig, zSig, &term0, &term1);
763                 sub128(aSig, 0, term0, term1, &rem0, &rem1);
764                 while ((sbits64) rem0 < 0) {
765                         --zSig;
766                         add128(rem0, rem1, 0, bSig, &rem0, &rem1);
767                 }
768                 zSig |= (rem1 != 0);
769         }
770         return roundAndPackFloat64(zSign, zExp, zSig);
771 
772 }
773 
774 float32 float32_div(float32 a, float32 b)
775 {
776         flag aSign, bSign, zSign;
777         int16 aExp, bExp, zExp;
778         bits32 aSig, bSig;
779         uint64_t zSig;
780 
781         aSig = extractFloat32Frac(a);
782         aExp = extractFloat32Exp(a);
783         aSign = extractFloat32Sign(a);
784         bSig = extractFloat32Frac(b);
785         bExp = extractFloat32Exp(b);
786         bSign = extractFloat32Sign(b);
787         zSign = aSign ^ bSign;
788         if (aExp == 0xFF) {
789                 if (bExp == 0xFF) {
790                 }
791                 return packFloat32(zSign, 0xFF, 0);
792         }
793         if (bExp == 0xFF) {
794                 return packFloat32(zSign, 0, 0);
795         }
796         if (bExp == 0) {
797                 if (bSig == 0) {
798                         return packFloat32(zSign, 0xFF, 0);
799                 }
800                 normalizeFloat32Subnormal(bSig, &bExp, &bSig);
801         }
802         if (aExp == 0) {
803                 if (aSig == 0)
804                         return packFloat32(zSign, 0, 0);
805                 normalizeFloat32Subnormal(aSig, &aExp, &aSig);
806         }
807         zExp = aExp - bExp + 0x7D;
808         aSig = (aSig | 0x00800000) << 7;
809         bSig = (bSig | 0x00800000) << 8;
810         if (bSig <= (aSig + aSig)) {
811                 aSig >>= 1;
812                 ++zExp;
813         }
814         zSig = (((bits64) aSig) << 32);
815         do_div(zSig, bSig);
816 
817         if ((zSig & 0x3F) == 0) {
818                 zSig |= (((bits64) bSig) * zSig != ((bits64) aSig) << 32);
819         }
820         return roundAndPackFloat32(zSign, zExp, (bits32)zSig);
821 
822 }
823 
824 float32 float32_mul(float32 a, float32 b)
825 {
826         char aSign, bSign, zSign;
827         int aExp, bExp, zExp;
828         unsigned int aSig, bSig;
829         unsigned long long zSig64;
830         unsigned int zSig;
831 
832         aSig = extractFloat32Frac(a);
833         aExp = extractFloat32Exp(a);
834         aSign = extractFloat32Sign(a);
835         bSig = extractFloat32Frac(b);
836         bExp = extractFloat32Exp(b);
837         bSign = extractFloat32Sign(b);
838         zSign = aSign ^ bSign;
839         if (aExp == 0) {
840                 if (aSig == 0)
841                         return packFloat32(zSign, 0, 0);
842                 normalizeFloat32Subnormal(aSig, &aExp, &aSig);
843         }
844         if (bExp == 0) {
845                 if (bSig == 0)
846                         return packFloat32(zSign, 0, 0);
847                 normalizeFloat32Subnormal(bSig, &bExp, &bSig);
848         }
849         if ((bExp == 0xff && bSig == 0) || (aExp == 0xff && aSig == 0))
850                 return roundAndPackFloat32(zSign, 0xff, 0);
851 
852         zExp = aExp + bExp - 0x7F;
853         aSig = (aSig | 0x00800000) << 7;
854         bSig = (bSig | 0x00800000) << 8;
855         shift64RightJamming(((unsigned long long)aSig) * bSig, 32, &zSig64);
856         zSig = zSig64;
857         if (0 <= (signed int)(zSig << 1)) {
858                 zSig <<= 1;
859                 --zExp;
860         }
861         return roundAndPackFloat32(zSign, zExp, zSig);
862 
863 }
864 
865 float64 float64_mul(float64 a, float64 b)
866 {
867         char aSign, bSign, zSign;
868         int aExp, bExp, zExp;
869         unsigned long long int aSig, bSig, zSig0, zSig1;
870 
871         aSig = extractFloat64Frac(a);
872         aExp = extractFloat64Exp(a);
873         aSign = extractFloat64Sign(a);
874         bSig = extractFloat64Frac(b);
875         bExp = extractFloat64Exp(b);
876         bSign = extractFloat64Sign(b);
877         zSign = aSign ^ bSign;
878 
879         if (aExp == 0) {
880                 if (aSig == 0)
881                         return packFloat64(zSign, 0, 0);
882                 normalizeFloat64Subnormal(aSig, &aExp, &aSig);
883         }
884         if (bExp == 0) {
885                 if (bSig == 0)
886                         return packFloat64(zSign, 0, 0);
887                 normalizeFloat64Subnormal(bSig, &bExp, &bSig);
888         }
889         if ((aExp == 0x7ff && aSig == 0) || (bExp == 0x7ff && bSig == 0))
890                 return roundAndPackFloat64(zSign, 0x7ff, 0);
891 
892         zExp = aExp + bExp - 0x3FF;
893         aSig = (aSig | 0x0010000000000000LL) << 10;
894         bSig = (bSig | 0x0010000000000000LL) << 11;
895         mul64To128(aSig, bSig, &zSig0, &zSig1);
896         zSig0 |= (zSig1 != 0);
897         if (0 <= (signed long long int)(zSig0 << 1)) {
898                 zSig0 <<= 1;
899                 --zExp;
900         }
901         return roundAndPackFloat64(zSign, zExp, zSig0);
902 }
903 
904 /*
905  * -------------------------------------------------------------------------------
906  *  Returns the result of converting the double-precision floating-point value
907  *  `a' to the single-precision floating-point format.  The conversion is
908  *  performed according to the IEC/IEEE Standard for Binary Floating-point
909  *  Arithmetic.
910  *  -------------------------------------------------------------------------------
911  *  */
912 float32 float64_to_float32(float64 a)
913 {
914     flag aSign;
915     int16 aExp;
916     bits64 aSig;
917     bits32 zSig;
918 
919     aSig = extractFloat64Frac( a );
920     aExp = extractFloat64Exp( a );
921     aSign = extractFloat64Sign( a );
922 
923     shift64RightJamming( aSig, 22, &aSig );
924     zSig = aSig;
925     if ( aExp || zSig ) {
926         zSig |= 0x40000000;
927         aExp -= 0x381;
928     }
929     return roundAndPackFloat32(aSign, aExp, zSig);
930 }
931 

~ [ source navigation ] ~ [ diff markup ] ~ [ identifier search ] ~

kernel.org | git.kernel.org | LWN.net | Project Home | SVN repository | Mail admin

Linux® is a registered trademark of Linus Torvalds in the United States and other countries.
TOMOYO® is a registered trademark of NTT DATA CORPORATION.

sflogo.php