URL
https://opencores.org/ocsvn/openrisc/openrisc/trunk
Subversion Repositories openrisc
[/] [openrisc/] [trunk/] [gnu-stable/] [gcc-4.5.1/] [libgcc/] [config/] [libbid/] [bid_internal.h] - Rev 816
Go to most recent revision | Compare with Previous | Blame | View Log
/* Copyright (C) 2007, 2009 Free Software Foundation, Inc. This file is part of GCC. GCC is free software; you can redistribute it and/or modify it under the terms of the GNU General Public License as published by the Free Software Foundation; either version 3, or (at your option) any later version. GCC is distributed in the hope that it will be useful, but WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more details. Under Section 7 of GPL version 3, you are granted additional permissions described in the GCC Runtime Library Exception, version 3.1, as published by the Free Software Foundation. You should have received a copy of the GNU General Public License and a copy of the GCC Runtime Library Exception along with this program; see the files COPYING3 and COPYING.RUNTIME respectively. If not, see <http://www.gnu.org/licenses/>. */ #ifndef __BIDECIMAL_H #define __BIDECIMAL_H #include "bid_conf.h" #include "bid_functions.h" #define __BID_INLINE__ static __inline /********************************************************************* * * Logical Shift Macros * *********************************************************************/ #define __shr_128(Q, A, k) \ { \ (Q).w[0] = (A).w[0] >> k; \ (Q).w[0] |= (A).w[1] << (64-k); \ (Q).w[1] = (A).w[1] >> k; \ } #define __shr_128_long(Q, A, k) \ { \ if((k)<64) { \ (Q).w[0] = (A).w[0] >> k; \ (Q).w[0] |= (A).w[1] << (64-k); \ (Q).w[1] = (A).w[1] >> k; \ } \ else { \ (Q).w[0] = (A).w[1]>>((k)-64); \ (Q).w[1] = 0; \ } \ } #define __shl_128_long(Q, A, k) \ { \ if((k)<64) { \ (Q).w[1] = (A).w[1] << k; \ (Q).w[1] |= (A).w[0] >> (64-k); \ (Q).w[0] = (A).w[0] << k; \ } \ else { \ (Q).w[1] = (A).w[0]<<((k)-64); \ (Q).w[0] = 0; \ } \ } #define __low_64(Q) (Q).w[0] /********************************************************************* * * String Macros * *********************************************************************/ #define tolower_macro(x) (((unsigned char)((x)-'A')<=('Z'-'A'))?((x)-'A'+'a'):(x)) /********************************************************************* * * Compare Macros * *********************************************************************/ // greater than // return 0 if A<=B // non-zero if A>B #define __unsigned_compare_gt_128(A, B) \ ((A.w[1]>B.w[1]) || ((A.w[1]==B.w[1]) && (A.w[0]>B.w[0]))) // greater-or-equal #define __unsigned_compare_ge_128(A, B) \ ((A.w[1]>B.w[1]) || ((A.w[1]==B.w[1]) && (A.w[0]>=B.w[0]))) #define __test_equal_128(A, B) (((A).w[1]==(B).w[1]) && ((A).w[0]==(B).w[0])) // tighten exponent range #define __tight_bin_range_128(bp, P, bin_expon) \ { \ UINT64 M; \ M = 1; \ (bp) = (bin_expon); \ if((bp)<63) { \ M <<= ((bp)+1); \ if((P).w[0] >= M) (bp)++; } \ else if((bp)>64) { \ M <<= ((bp)+1-64); \ if(((P).w[1]>M) ||((P).w[1]==M && (P).w[0]))\ (bp)++; } \ else if((P).w[1]) (bp)++; \ } /********************************************************************* * * Add/Subtract Macros * *********************************************************************/ // add 64-bit value to 128-bit #define __add_128_64(R128, A128, B64) \ { \ UINT64 R64H; \ R64H = (A128).w[1]; \ (R128).w[0] = (B64) + (A128).w[0]; \ if((R128).w[0] < (B64)) \ R64H ++; \ (R128).w[1] = R64H; \ } // subtract 64-bit value from 128-bit #define __sub_128_64(R128, A128, B64) \ { \ UINT64 R64H; \ R64H = (A128).w[1]; \ if((A128).w[0] < (B64)) \ R64H --; \ (R128).w[1] = R64H; \ (R128).w[0] = (A128).w[0] - (B64); \ } // add 128-bit value to 128-bit // assume no carry-out #define __add_128_128(R128, A128, B128) \ { \ UINT128 Q128; \ Q128.w[1] = (A128).w[1]+(B128).w[1]; \ Q128.w[0] = (B128).w[0] + (A128).w[0]; \ if(Q128.w[0] < (B128).w[0]) \ Q128.w[1] ++; \ (R128).w[1] = Q128.w[1]; \ (R128).w[0] = Q128.w[0]; \ } #define __sub_128_128(R128, A128, B128) \ { \ UINT128 Q128; \ Q128.w[1] = (A128).w[1]-(B128).w[1]; \ Q128.w[0] = (A128).w[0] - (B128).w[0]; \ if((A128).w[0] < (B128).w[0]) \ Q128.w[1] --; \ (R128).w[1] = Q128.w[1]; \ (R128).w[0] = Q128.w[0]; \ } #define __add_carry_out(S, CY, X, Y) \ { \ UINT64 X1=X; \ S = X + Y; \ CY = (S<X1) ? 1 : 0; \ } #define __add_carry_in_out(S, CY, X, Y, CI) \ { \ UINT64 X1; \ X1 = X + CI; \ S = X1 + Y; \ CY = ((S<X1) || (X1<CI)) ? 1 : 0; \ } #define __sub_borrow_out(S, CY, X, Y) \ { \ UINT64 X1=X; \ S = X - Y; \ CY = (S>X1) ? 1 : 0; \ } #define __sub_borrow_in_out(S, CY, X, Y, CI) \ { \ UINT64 X1, X0=X; \ X1 = X - CI; \ S = X1 - Y; \ CY = ((S>X1) || (X1>X0)) ? 1 : 0; \ } // increment C128 and check for rounding overflow: // if (C_128) = 10^34 then (C_128) = 10^33 and increment the exponent #define INCREMENT(C_128, exp) \ { \ C_128.w[0]++; \ if (C_128.w[0] == 0) C_128.w[1]++; \ if (C_128.w[1] == 0x0001ed09bead87c0ull && \ C_128.w[0] == 0x378d8e6400000000ull) { \ exp++; \ C_128.w[1] = 0x0000314dc6448d93ull; \ C_128.w[0] = 0x38c15b0a00000000ull; \ } \ } // decrement C128 and check for rounding underflow, but only at the // boundary: if C_128 = 10^33 - 1 and exp > 0 then C_128 = 10^34 - 1 // and decrement the exponent #define DECREMENT(C_128, exp) \ { \ C_128.w[0]--; \ if (C_128.w[0] == 0xffffffffffffffffull) C_128.w[1]--; \ if (C_128.w[1] == 0x0000314dc6448d93ull && \ C_128.w[0] == 0x38c15b09ffffffffull && exp > 0) { \ exp--; \ C_128.w[1] = 0x0001ed09bead87c0ull; \ C_128.w[0] = 0x378d8e63ffffffffull; \ } \ } /********************************************************************* * * Multiply Macros * *********************************************************************/ #define __mul_64x64_to_64(P64, CX, CY) (P64) = (CX) * (CY) /*************************************** * Signed, Full 64x64-bit Multiply ***************************************/ #define __imul_64x64_to_128(P, CX, CY) \ { \ UINT64 SX, SY; \ __mul_64x64_to_128(P, CX, CY); \ \ SX = ((SINT64)(CX))>>63; \ SY = ((SINT64)(CY))>>63; \ SX &= CY; SY &= CX; \ \ (P).w[1] = (P).w[1] - SX - SY; \ } /*************************************** * Signed, Full 64x128-bit Multiply ***************************************/ #define __imul_64x128_full(Ph, Ql, A, B) \ { \ UINT128 ALBL, ALBH, QM2, QM; \ \ __imul_64x64_to_128(ALBH, (A), (B).w[1]); \ __imul_64x64_to_128(ALBL, (A), (B).w[0]); \ \ (Ql).w[0] = ALBL.w[0]; \ QM.w[0] = ALBL.w[1]; \ QM.w[1] = ((SINT64)ALBL.w[1])>>63; \ __add_128_128(QM2, ALBH, QM); \ (Ql).w[1] = QM2.w[0]; \ Ph = QM2.w[1]; \ } /***************************************************** * Unsigned Multiply Macros *****************************************************/ // get full 64x64bit product // #define __mul_64x64_to_128(P, CX, CY) \ { \ UINT64 CXH, CXL, CYH,CYL,PL,PH,PM,PM2;\ CXH = (CX) >> 32; \ CXL = (UINT32)(CX); \ CYH = (CY) >> 32; \ CYL = (UINT32)(CY); \ \ PM = CXH*CYL; \ PH = CXH*CYH; \ PL = CXL*CYL; \ PM2 = CXL*CYH; \ PH += (PM>>32); \ PM = (UINT64)((UINT32)PM)+PM2+(PL>>32); \ \ (P).w[1] = PH + (PM>>32); \ (P).w[0] = (PM<<32)+(UINT32)PL; \ } // get full 64x64bit product // Note: // This macro is used for CX < 2^61, CY < 2^61 // #define __mul_64x64_to_128_fast(P, CX, CY) \ { \ UINT64 CXH, CXL, CYH, CYL, PL, PH, PM; \ CXH = (CX) >> 32; \ CXL = (UINT32)(CX); \ CYH = (CY) >> 32; \ CYL = (UINT32)(CY); \ \ PM = CXH*CYL; \ PL = CXL*CYL; \ PH = CXH*CYH; \ PM += CXL*CYH; \ PM += (PL>>32); \ \ (P).w[1] = PH + (PM>>32); \ (P).w[0] = (PM<<32)+(UINT32)PL; \ } // used for CX< 2^60 #define __sqr64_fast(P, CX) \ { \ UINT64 CXH, CXL, PL, PH, PM; \ CXH = (CX) >> 32; \ CXL = (UINT32)(CX); \ \ PM = CXH*CXL; \ PL = CXL*CXL; \ PH = CXH*CXH; \ PM += PM; \ PM += (PL>>32); \ \ (P).w[1] = PH + (PM>>32); \ (P).w[0] = (PM<<32)+(UINT32)PL; \ } // get full 64x64bit product // Note: // This implementation is used for CX < 2^61, CY < 2^61 // #define __mul_64x64_to_64_high_fast(P, CX, CY) \ { \ UINT64 CXH, CXL, CYH, CYL, PL, PH, PM; \ CXH = (CX) >> 32; \ CXL = (UINT32)(CX); \ CYH = (CY) >> 32; \ CYL = (UINT32)(CY); \ \ PM = CXH*CYL; \ PL = CXL*CYL; \ PH = CXH*CYH; \ PM += CXL*CYH; \ PM += (PL>>32); \ \ (P) = PH + (PM>>32); \ } // get full 64x64bit product // #define __mul_64x64_to_128_full(P, CX, CY) \ { \ UINT64 CXH, CXL, CYH,CYL,PL,PH,PM,PM2;\ CXH = (CX) >> 32; \ CXL = (UINT32)(CX); \ CYH = (CY) >> 32; \ CYL = (UINT32)(CY); \ \ PM = CXH*CYL; \ PH = CXH*CYH; \ PL = CXL*CYL; \ PM2 = CXL*CYH; \ PH += (PM>>32); \ PM = (UINT64)((UINT32)PM)+PM2+(PL>>32); \ \ (P).w[1] = PH + (PM>>32); \ (P).w[0] = (PM<<32)+(UINT32)PL; \ } #define __mul_128x128_high(Q, A, B) \ { \ UINT128 ALBL, ALBH, AHBL, AHBH, QM, QM2; \ \ __mul_64x64_to_128(ALBH, (A).w[0], (B).w[1]); \ __mul_64x64_to_128(AHBL, (B).w[0], (A).w[1]); \ __mul_64x64_to_128(ALBL, (A).w[0], (B).w[0]); \ __mul_64x64_to_128(AHBH, (A).w[1],(B).w[1]); \ \ __add_128_128(QM, ALBH, AHBL); \ __add_128_64(QM2, QM, ALBL.w[1]); \ __add_128_64((Q), AHBH, QM2.w[1]); \ } #define __mul_128x128_full(Qh, Ql, A, B) \ { \ UINT128 ALBL, ALBH, AHBL, AHBH, QM, QM2; \ \ __mul_64x64_to_128(ALBH, (A).w[0], (B).w[1]); \ __mul_64x64_to_128(AHBL, (B).w[0], (A).w[1]); \ __mul_64x64_to_128(ALBL, (A).w[0], (B).w[0]); \ __mul_64x64_to_128(AHBH, (A).w[1],(B).w[1]); \ \ __add_128_128(QM, ALBH, AHBL); \ (Ql).w[0] = ALBL.w[0]; \ __add_128_64(QM2, QM, ALBL.w[1]); \ __add_128_64((Qh), AHBH, QM2.w[1]); \ (Ql).w[1] = QM2.w[0]; \ } #define __mul_128x128_low(Ql, A, B) \ { \ UINT128 ALBL; \ UINT64 QM64; \ \ __mul_64x64_to_128(ALBL, (A).w[0], (B).w[0]); \ QM64 = (B).w[0]*(A).w[1] + (A).w[0]*(B).w[1]; \ \ (Ql).w[0] = ALBL.w[0]; \ (Ql).w[1] = QM64 + ALBL.w[1]; \ } #define __mul_64x128_low(Ql, A, B) \ { \ UINT128 ALBL, ALBH, QM2; \ __mul_64x64_to_128(ALBH, (A), (B).w[1]); \ __mul_64x64_to_128(ALBL, (A), (B).w[0]); \ (Ql).w[0] = ALBL.w[0]; \ __add_128_64(QM2, ALBH, ALBL.w[1]); \ (Ql).w[1] = QM2.w[0]; \ } #define __mul_64x128_full(Ph, Ql, A, B) \ { \ UINT128 ALBL, ALBH, QM2; \ \ __mul_64x64_to_128(ALBH, (A), (B).w[1]); \ __mul_64x64_to_128(ALBL, (A), (B).w[0]); \ \ (Ql).w[0] = ALBL.w[0]; \ __add_128_64(QM2, ALBH, ALBL.w[1]); \ (Ql).w[1] = QM2.w[0]; \ Ph = QM2.w[1]; \ } #define __mul_64x128_to_192(Q, A, B) \ { \ UINT128 ALBL, ALBH, QM2; \ \ __mul_64x64_to_128(ALBH, (A), (B).w[1]); \ __mul_64x64_to_128(ALBL, (A), (B).w[0]); \ \ (Q).w[0] = ALBL.w[0]; \ __add_128_64(QM2, ALBH, ALBL.w[1]); \ (Q).w[1] = QM2.w[0]; \ (Q).w[2] = QM2.w[1]; \ } #define __mul_64x128_to192(Q, A, B) \ { \ UINT128 ALBL, ALBH, QM2; \ \ __mul_64x64_to_128(ALBH, (A), (B).w[1]); \ __mul_64x64_to_128(ALBL, (A), (B).w[0]); \ \ (Q).w[0] = ALBL.w[0]; \ __add_128_64(QM2, ALBH, ALBL.w[1]); \ (Q).w[1] = QM2.w[0]; \ (Q).w[2] = QM2.w[1]; \ } #define __mul_128x128_to_256(P256, A, B) \ { \ UINT128 Qll, Qlh; \ UINT64 Phl, Phh, CY1, CY2; \ \ __mul_64x128_full(Phl, Qll, A.w[0], B); \ __mul_64x128_full(Phh, Qlh, A.w[1], B); \ (P256).w[0] = Qll.w[0]; \ __add_carry_out((P256).w[1],CY1, Qlh.w[0], Qll.w[1]); \ __add_carry_in_out((P256).w[2],CY2, Qlh.w[1], Phl, CY1); \ (P256).w[3] = Phh + CY2; \ } // // For better performance, will check A.w[1] against 0, // but not B.w[1] // Use this macro accordingly #define __mul_128x128_to_256_check_A(P256, A, B) \ { \ UINT128 Qll, Qlh; \ UINT64 Phl, Phh, CY1, CY2; \ \ __mul_64x128_full(Phl, Qll, A.w[0], B); \ (P256).w[0] = Qll.w[0]; \ if(A.w[1]) { \ __mul_64x128_full(Phh, Qlh, A.w[1], B); \ __add_carry_out((P256).w[1],CY1, Qlh.w[0], Qll.w[1]); \ __add_carry_in_out((P256).w[2],CY2, Qlh.w[1], Phl, CY1); \ (P256).w[3] = Phh + CY2; } \ else { \ (P256).w[1] = Qll.w[1]; \ (P256).w[2] = Phl; \ (P256).w[3] = 0; } \ } #define __mul_64x192_to_256(lP, lA, lB) \ { \ UINT128 lP0,lP1,lP2; \ UINT64 lC; \ __mul_64x64_to_128(lP0, lA, (lB).w[0]); \ __mul_64x64_to_128(lP1, lA, (lB).w[1]); \ __mul_64x64_to_128(lP2, lA, (lB).w[2]); \ (lP).w[0] = lP0.w[0]; \ __add_carry_out((lP).w[1],lC,lP1.w[0],lP0.w[1]); \ __add_carry_in_out((lP).w[2],lC,lP2.w[0],lP1.w[1],lC); \ (lP).w[3] = lP2.w[1] + lC; \ } #define __mul_64x256_to_320(P, A, B) \ { \ UINT128 lP0,lP1,lP2,lP3; \ UINT64 lC; \ __mul_64x64_to_128(lP0, A, (B).w[0]); \ __mul_64x64_to_128(lP1, A, (B).w[1]); \ __mul_64x64_to_128(lP2, A, (B).w[2]); \ __mul_64x64_to_128(lP3, A, (B).w[3]); \ (P).w[0] = lP0.w[0]; \ __add_carry_out((P).w[1],lC,lP1.w[0],lP0.w[1]); \ __add_carry_in_out((P).w[2],lC,lP2.w[0],lP1.w[1],lC); \ __add_carry_in_out((P).w[3],lC,lP3.w[0],lP2.w[1],lC); \ (P).w[4] = lP3.w[1] + lC; \ } #define __mul_192x192_to_384(P, A, B) \ { \ UINT256 P0,P1,P2; \ UINT64 CY; \ __mul_64x192_to_256(P0, (A).w[0], B); \ __mul_64x192_to_256(P1, (A).w[1], B); \ __mul_64x192_to_256(P2, (A).w[2], B); \ (P).w[0] = P0.w[0]; \ __add_carry_out((P).w[1],CY,P1.w[0],P0.w[1]); \ __add_carry_in_out((P).w[2],CY,P1.w[1],P0.w[2],CY); \ __add_carry_in_out((P).w[3],CY,P1.w[2],P0.w[3],CY); \ (P).w[4] = P1.w[3] + CY; \ __add_carry_out((P).w[2],CY,P2.w[0],(P).w[2]); \ __add_carry_in_out((P).w[3],CY,P2.w[1],(P).w[3],CY); \ __add_carry_in_out((P).w[4],CY,P2.w[2],(P).w[4],CY); \ (P).w[5] = P2.w[3] + CY; \ } #define __mul_64x320_to_384(P, A, B) \ { \ UINT128 lP0,lP1,lP2,lP3,lP4; \ UINT64 lC; \ __mul_64x64_to_128(lP0, A, (B).w[0]); \ __mul_64x64_to_128(lP1, A, (B).w[1]); \ __mul_64x64_to_128(lP2, A, (B).w[2]); \ __mul_64x64_to_128(lP3, A, (B).w[3]); \ __mul_64x64_to_128(lP4, A, (B).w[4]); \ (P).w[0] = lP0.w[0]; \ __add_carry_out((P).w[1],lC,lP1.w[0],lP0.w[1]); \ __add_carry_in_out((P).w[2],lC,lP2.w[0],lP1.w[1],lC); \ __add_carry_in_out((P).w[3],lC,lP3.w[0],lP2.w[1],lC); \ __add_carry_in_out((P).w[4],lC,lP4.w[0],lP3.w[1],lC); \ (P).w[5] = lP4.w[1] + lC; \ } // A*A // Full 128x128-bit product #define __sqr128_to_256(P256, A) \ { \ UINT128 Qll, Qlh, Qhh; \ UINT64 TMP_C1, TMP_C2; \ \ __mul_64x64_to_128(Qhh, A.w[1], A.w[1]); \ __mul_64x64_to_128(Qlh, A.w[0], A.w[1]); \ Qhh.w[1] += (Qlh.w[1]>>63); \ Qlh.w[1] = (Qlh.w[1]+Qlh.w[1])|(Qlh.w[0]>>63); \ Qlh.w[0] += Qlh.w[0]; \ __mul_64x64_to_128(Qll, A.w[0], A.w[0]); \ \ __add_carry_out((P256).w[1],TMP_C1, Qlh.w[0], Qll.w[1]); \ (P256).w[0] = Qll.w[0]; \ __add_carry_in_out((P256).w[2],TMP_C2, Qlh.w[1], Qhh.w[0], TMP_C1); \ (P256).w[3] = Qhh.w[1]+TMP_C2; \ } #define __mul_128x128_to_256_low_high(PQh, PQl, A, B) \ { \ UINT128 Qll, Qlh; \ UINT64 Phl, Phh, C1, C2; \ \ __mul_64x128_full(Phl, Qll, A.w[0], B); \ __mul_64x128_full(Phh, Qlh, A.w[1], B); \ (PQl).w[0] = Qll.w[0]; \ __add_carry_out((PQl).w[1],C1, Qlh.w[0], Qll.w[1]); \ __add_carry_in_out((PQh).w[0],C2, Qlh.w[1], Phl, C1); \ (PQh).w[1] = Phh + C2; \ } #define __mul_256x256_to_512(P, A, B) \ { \ UINT512 P0,P1,P2,P3; \ UINT64 CY; \ __mul_64x256_to_320(P0, (A).w[0], B); \ __mul_64x256_to_320(P1, (A).w[1], B); \ __mul_64x256_to_320(P2, (A).w[2], B); \ __mul_64x256_to_320(P3, (A).w[3], B); \ (P).w[0] = P0.w[0]; \ __add_carry_out((P).w[1],CY,P1.w[0],P0.w[1]); \ __add_carry_in_out((P).w[2],CY,P1.w[1],P0.w[2],CY); \ __add_carry_in_out((P).w[3],CY,P1.w[2],P0.w[3],CY); \ __add_carry_in_out((P).w[4],CY,P1.w[3],P0.w[4],CY); \ (P).w[5] = P1.w[4] + CY; \ __add_carry_out((P).w[2],CY,P2.w[0],(P).w[2]); \ __add_carry_in_out((P).w[3],CY,P2.w[1],(P).w[3],CY); \ __add_carry_in_out((P).w[4],CY,P2.w[2],(P).w[4],CY); \ __add_carry_in_out((P).w[5],CY,P2.w[3],(P).w[5],CY); \ (P).w[6] = P2.w[4] + CY; \ __add_carry_out((P).w[3],CY,P3.w[0],(P).w[3]); \ __add_carry_in_out((P).w[4],CY,P3.w[1],(P).w[4],CY); \ __add_carry_in_out((P).w[5],CY,P3.w[2],(P).w[5],CY); \ __add_carry_in_out((P).w[6],CY,P3.w[3],(P).w[6],CY); \ (P).w[7] = P3.w[4] + CY; \ } #define __mul_192x256_to_448(P, A, B) \ { \ UINT512 P0,P1,P2; \ UINT64 CY; \ __mul_64x256_to_320(P0, (A).w[0], B); \ __mul_64x256_to_320(P1, (A).w[1], B); \ __mul_64x256_to_320(P2, (A).w[2], B); \ (P).w[0] = P0.w[0]; \ __add_carry_out((P).w[1],CY,P1.w[0],P0.w[1]); \ __add_carry_in_out((P).w[2],CY,P1.w[1],P0.w[2],CY); \ __add_carry_in_out((P).w[3],CY,P1.w[2],P0.w[3],CY); \ __add_carry_in_out((P).w[4],CY,P1.w[3],P0.w[4],CY); \ (P).w[5] = P1.w[4] + CY; \ __add_carry_out((P).w[2],CY,P2.w[0],(P).w[2]); \ __add_carry_in_out((P).w[3],CY,P2.w[1],(P).w[3],CY); \ __add_carry_in_out((P).w[4],CY,P2.w[2],(P).w[4],CY); \ __add_carry_in_out((P).w[5],CY,P2.w[3],(P).w[5],CY); \ (P).w[6] = P2.w[4] + CY; \ } #define __mul_320x320_to_640(P, A, B) \ { \ UINT512 P0,P1,P2,P3; \ UINT64 CY; \ __mul_256x256_to_512((P), (A), B); \ __mul_64x256_to_320(P1, (A).w[4], B); \ __mul_64x256_to_320(P2, (B).w[4], A); \ __mul_64x64_to_128(P3, (A).w[4], (B).w[4]); \ __add_carry_out((P0).w[0],CY,P1.w[0],P2.w[0]); \ __add_carry_in_out((P0).w[1],CY,P1.w[1],P2.w[1],CY); \ __add_carry_in_out((P0).w[2],CY,P1.w[2],P2.w[2],CY); \ __add_carry_in_out((P0).w[3],CY,P1.w[3],P2.w[3],CY); \ __add_carry_in_out((P0).w[4],CY,P1.w[4],P2.w[4],CY); \ P3.w[1] += CY; \ __add_carry_out((P).w[4],CY,(P).w[4],P0.w[0]); \ __add_carry_in_out((P).w[5],CY,(P).w[5],P0.w[1],CY); \ __add_carry_in_out((P).w[6],CY,(P).w[6],P0.w[2],CY); \ __add_carry_in_out((P).w[7],CY,(P).w[7],P0.w[3],CY); \ __add_carry_in_out((P).w[8],CY,P3.w[0],P0.w[4],CY); \ (P).w[9] = P3.w[1] + CY; \ } #define __mul_384x384_to_768(P, A, B) \ { \ UINT512 P0,P1,P2,P3; \ UINT64 CY; \ __mul_320x320_to_640((P), (A), B); \ __mul_64x320_to_384(P1, (A).w[5], B); \ __mul_64x320_to_384(P2, (B).w[5], A); \ __mul_64x64_to_128(P3, (A).w[5], (B).w[5]); \ __add_carry_out((P0).w[0],CY,P1.w[0],P2.w[0]); \ __add_carry_in_out((P0).w[1],CY,P1.w[1],P2.w[1],CY); \ __add_carry_in_out((P0).w[2],CY,P1.w[2],P2.w[2],CY); \ __add_carry_in_out((P0).w[3],CY,P1.w[3],P2.w[3],CY); \ __add_carry_in_out((P0).w[4],CY,P1.w[4],P2.w[4],CY); \ __add_carry_in_out((P0).w[5],CY,P1.w[5],P2.w[5],CY); \ P3.w[1] += CY; \ __add_carry_out((P).w[5],CY,(P).w[5],P0.w[0]); \ __add_carry_in_out((P).w[6],CY,(P).w[6],P0.w[1],CY); \ __add_carry_in_out((P).w[7],CY,(P).w[7],P0.w[2],CY); \ __add_carry_in_out((P).w[8],CY,(P).w[8],P0.w[3],CY); \ __add_carry_in_out((P).w[9],CY,(P).w[9],P0.w[4],CY); \ __add_carry_in_out((P).w[10],CY,P3.w[0],P0.w[5],CY); \ (P).w[11] = P3.w[1] + CY; \ } #define __mul_64x128_short(Ql, A, B) \ { \ UINT64 ALBH_L; \ \ __mul_64x64_to_64(ALBH_L, (A),(B).w[1]); \ __mul_64x64_to_128((Ql), (A), (B).w[0]); \ \ (Ql).w[1] += ALBH_L; \ } #define __scale128_10(D,_TMP) \ { \ UINT128 _TMP2,_TMP8; \ _TMP2.w[1] = (_TMP.w[1]<<1)|(_TMP.w[0]>>63); \ _TMP2.w[0] = _TMP.w[0]<<1; \ _TMP8.w[1] = (_TMP.w[1]<<3)|(_TMP.w[0]>>61); \ _TMP8.w[0] = _TMP.w[0]<<3; \ __add_128_128(D, _TMP2, _TMP8); \ } // 64x64-bit product #define __mul_64x64_to_128MACH(P128, CX64, CY64) \ { \ UINT64 CXH,CXL,CYH,CYL,PL,PH,PM,PM2; \ CXH = (CX64) >> 32; \ CXL = (UINT32)(CX64); \ CYH = (CY64) >> 32; \ CYL = (UINT32)(CY64); \ PM = CXH*CYL; \ PH = CXH*CYH; \ PL = CXL*CYL; \ PM2 = CXL*CYH; \ PH += (PM>>32); \ PM = (UINT64)((UINT32)PM)+PM2+(PL>>32); \ (P128).w[1] = PH + (PM>>32); \ (P128).w[0] = (PM<<32)+(UINT32)PL; \ } // 64x64-bit product #define __mul_64x64_to_128HIGH(P64, CX64, CY64) \ { \ UINT64 CXH,CXL,CYH,CYL,PL,PH,PM,PM2; \ CXH = (CX64) >> 32; \ CXL = (UINT32)(CX64); \ CYH = (CY64) >> 32; \ CYL = (UINT32)(CY64); \ PM = CXH*CYL; \ PH = CXH*CYH; \ PL = CXL*CYL; \ PM2 = CXL*CYH; \ PH += (PM>>32); \ PM = (UINT64)((UINT32)PM)+PM2+(PL>>32); \ P64 = PH + (PM>>32); \ } #define __mul_128x64_to_128(Q128, A64, B128) \ { \ UINT64 ALBH_L; \ ALBH_L = (A64) * (B128).w[1]; \ __mul_64x64_to_128MACH((Q128), (A64), (B128).w[0]); \ (Q128).w[1] += ALBH_L; \ } // might simplify by calculating just QM2.w[0] #define __mul_64x128_to_128(Ql, A, B) \ { \ UINT128 ALBL, ALBH, QM2; \ __mul_64x64_to_128(ALBH, (A), (B).w[1]); \ __mul_64x64_to_128(ALBL, (A), (B).w[0]); \ (Ql).w[0] = ALBL.w[0]; \ __add_128_64(QM2, ALBH, ALBL.w[1]); \ (Ql).w[1] = QM2.w[0]; \ } /********************************************************************* * * BID Pack/Unpack Macros * *********************************************************************/ ///////////////////////////////////////// // BID64 definitions //////////////////////////////////////// #define DECIMAL_MAX_EXPON_64 767 #define DECIMAL_EXPONENT_BIAS 398 #define MAX_FORMAT_DIGITS 16 ///////////////////////////////////////// // BID128 definitions //////////////////////////////////////// #define DECIMAL_MAX_EXPON_128 12287 #define DECIMAL_EXPONENT_BIAS_128 6176 #define MAX_FORMAT_DIGITS_128 34 ///////////////////////////////////////// // BID32 definitions //////////////////////////////////////// #define DECIMAL_MAX_EXPON_32 191 #define DECIMAL_EXPONENT_BIAS_32 101 #define MAX_FORMAT_DIGITS_32 7 //////////////////////////////////////// // Constant Definitions /////////////////////////////////////// #define SPECIAL_ENCODING_MASK64 0x6000000000000000ull #define INFINITY_MASK64 0x7800000000000000ull #define SINFINITY_MASK64 0xf800000000000000ull #define SSNAN_MASK64 0xfc00000000000000ull #define NAN_MASK64 0x7c00000000000000ull #define SNAN_MASK64 0x7e00000000000000ull #define QUIET_MASK64 0xfdffffffffffffffull #define LARGE_COEFF_MASK64 0x0007ffffffffffffull #define LARGE_COEFF_HIGH_BIT64 0x0020000000000000ull #define SMALL_COEFF_MASK64 0x001fffffffffffffull #define EXPONENT_MASK64 0x3ff #define EXPONENT_SHIFT_LARGE64 51 #define EXPONENT_SHIFT_SMALL64 53 #define LARGEST_BID64 0x77fb86f26fc0ffffull #define SMALLEST_BID64 0xf7fb86f26fc0ffffull #define SMALL_COEFF_MASK128 0x0001ffffffffffffull #define LARGE_COEFF_MASK128 0x00007fffffffffffull #define EXPONENT_MASK128 0x3fff #define LARGEST_BID128_HIGH 0x5fffed09bead87c0ull #define LARGEST_BID128_LOW 0x378d8e63ffffffffull #define SPECIAL_ENCODING_MASK32 0x60000000ul #define INFINITY_MASK32 0x78000000ul #define LARGE_COEFF_MASK32 0x007ffffful #define LARGE_COEFF_HIGH_BIT32 0x00800000ul #define SMALL_COEFF_MASK32 0x001ffffful #define EXPONENT_MASK32 0xff #define LARGEST_BID32 0x77f8967f #define NAN_MASK32 0x7c000000 #define SNAN_MASK32 0x7e000000 #define MASK_BINARY_EXPONENT 0x7ff0000000000000ull #define BINARY_EXPONENT_BIAS 0x3ff #define UPPER_EXPON_LIMIT 51 // data needed for BID pack/unpack macros extern UINT64 round_const_table[][19]; extern UINT128 reciprocals10_128[]; extern int recip_scale[]; extern UINT128 power10_table_128[]; extern int estimate_decimal_digits[]; extern int estimate_bin_expon[]; extern UINT64 power10_index_binexp[]; extern int short_recip_scale[]; extern UINT64 reciprocals10_64[]; extern UINT128 power10_index_binexp_128[]; extern UINT128 round_const_table_128[][36]; ////////////////////////////////////////////// // Status Flag Handling ///////////////////////////////////////////// #define __set_status_flags(fpsc, status) *(fpsc) |= status #define is_inexact(fpsc) ((*(fpsc))&INEXACT_EXCEPTION) __BID_INLINE__ UINT64 unpack_BID64 (UINT64 * psign_x, int *pexponent_x, UINT64 * pcoefficient_x, UINT64 x) { UINT64 tmp, coeff; *psign_x = x & 0x8000000000000000ull; if ((x & SPECIAL_ENCODING_MASK64) == SPECIAL_ENCODING_MASK64) { // special encodings // coefficient coeff = (x & LARGE_COEFF_MASK64) | LARGE_COEFF_HIGH_BIT64; if ((x & INFINITY_MASK64) == INFINITY_MASK64) { *pexponent_x = 0; *pcoefficient_x = x & 0xfe03ffffffffffffull; if ((x & 0x0003ffffffffffffull) >= 1000000000000000ull) *pcoefficient_x = x & 0xfe00000000000000ull; if ((x & NAN_MASK64) == INFINITY_MASK64) *pcoefficient_x = x & SINFINITY_MASK64; return 0; // NaN or Infinity } // check for non-canonical values if (coeff >= 10000000000000000ull) coeff = 0; *pcoefficient_x = coeff; // get exponent tmp = x >> EXPONENT_SHIFT_LARGE64; *pexponent_x = (int) (tmp & EXPONENT_MASK64); return coeff; } // exponent tmp = x >> EXPONENT_SHIFT_SMALL64; *pexponent_x = (int) (tmp & EXPONENT_MASK64); // coefficient *pcoefficient_x = (x & SMALL_COEFF_MASK64); return *pcoefficient_x; } // // BID64 pack macro (general form) // __BID_INLINE__ UINT64 get_BID64 (UINT64 sgn, int expon, UINT64 coeff, int rmode, unsigned *fpsc) { UINT128 Stemp, Q_low; UINT64 QH, r, mask, C64, remainder_h, CY, carry; int extra_digits, amount, amount2; unsigned status; if (coeff > 9999999999999999ull) { expon++; coeff = 1000000000000000ull; } // check for possible underflow/overflow if (((unsigned) expon) >= 3 * 256) { if (expon < 0) { // underflow if (expon + MAX_FORMAT_DIGITS < 0) { #ifdef SET_STATUS_FLAGS __set_status_flags (fpsc, UNDERFLOW_EXCEPTION | INEXACT_EXCEPTION); #endif #ifndef IEEE_ROUND_NEAREST_TIES_AWAY #ifndef IEEE_ROUND_NEAREST if (rmode == ROUNDING_DOWN && sgn) return 0x8000000000000001ull; if (rmode == ROUNDING_UP && !sgn) return 1ull; #endif #endif // result is 0 return sgn; } #ifndef IEEE_ROUND_NEAREST_TIES_AWAY #ifndef IEEE_ROUND_NEAREST if (sgn && (unsigned) (rmode - 1) < 2) rmode = 3 - rmode; #endif #endif // get digits to be shifted out extra_digits = -expon; coeff += round_const_table[rmode][extra_digits]; // get coeff*(2^M[extra_digits])/10^extra_digits __mul_64x128_full (QH, Q_low, coeff, reciprocals10_128[extra_digits]); // now get P/10^extra_digits: shift Q_high right by M[extra_digits]-128 amount = recip_scale[extra_digits]; C64 = QH >> amount; #ifndef IEEE_ROUND_NEAREST_TIES_AWAY #ifndef IEEE_ROUND_NEAREST if (rmode == 0) //ROUNDING_TO_NEAREST #endif if (C64 & 1) { // check whether fractional part of initial_P/10^extra_digits is exactly .5 // get remainder amount2 = 64 - amount; remainder_h = 0; remainder_h--; remainder_h >>= amount2; remainder_h = remainder_h & QH; if (!remainder_h && (Q_low.w[1] < reciprocals10_128[extra_digits].w[1] || (Q_low.w[1] == reciprocals10_128[extra_digits].w[1] && Q_low.w[0] < reciprocals10_128[extra_digits].w[0]))) { C64--; } } #endif #ifdef SET_STATUS_FLAGS if (is_inexact (fpsc)) __set_status_flags (fpsc, UNDERFLOW_EXCEPTION); else { status = INEXACT_EXCEPTION; // get remainder remainder_h = QH << (64 - amount); switch (rmode) { case ROUNDING_TO_NEAREST: case ROUNDING_TIES_AWAY: // test whether fractional part is 0 if (remainder_h == 0x8000000000000000ull && (Q_low.w[1] < reciprocals10_128[extra_digits].w[1] || (Q_low.w[1] == reciprocals10_128[extra_digits].w[1] && Q_low.w[0] < reciprocals10_128[extra_digits].w[0]))) status = EXACT_STATUS; break; case ROUNDING_DOWN: case ROUNDING_TO_ZERO: if (!remainder_h && (Q_low.w[1] < reciprocals10_128[extra_digits].w[1] || (Q_low.w[1] == reciprocals10_128[extra_digits].w[1] && Q_low.w[0] < reciprocals10_128[extra_digits].w[0]))) status = EXACT_STATUS; break; default: // round up __add_carry_out (Stemp.w[0], CY, Q_low.w[0], reciprocals10_128[extra_digits].w[0]); __add_carry_in_out (Stemp.w[1], carry, Q_low.w[1], reciprocals10_128[extra_digits].w[1], CY); if ((remainder_h >> (64 - amount)) + carry >= (((UINT64) 1) << amount)) status = EXACT_STATUS; } if (status != EXACT_STATUS) __set_status_flags (fpsc, UNDERFLOW_EXCEPTION | status); } #endif return sgn | C64; } while (coeff < 1000000000000000ull && expon >= 3 * 256) { expon--; coeff = (coeff << 3) + (coeff << 1); } if (expon > DECIMAL_MAX_EXPON_64) { #ifdef SET_STATUS_FLAGS __set_status_flags (fpsc, OVERFLOW_EXCEPTION | INEXACT_EXCEPTION); #endif // overflow r = sgn | INFINITY_MASK64; switch (rmode) { case ROUNDING_DOWN: if (!sgn) r = LARGEST_BID64; break; case ROUNDING_TO_ZERO: r = sgn | LARGEST_BID64; break; case ROUNDING_UP: // round up if (sgn) r = SMALLEST_BID64; } return r; } } mask = 1; mask <<= EXPONENT_SHIFT_SMALL64; // check whether coefficient fits in 10*5+3 bits if (coeff < mask) { r = expon; r <<= EXPONENT_SHIFT_SMALL64; r |= (coeff | sgn); return r; } // special format // eliminate the case coeff==10^16 after rounding if (coeff == 10000000000000000ull) { r = expon + 1; r <<= EXPONENT_SHIFT_SMALL64; r |= (1000000000000000ull | sgn); return r; } r = expon; r <<= EXPONENT_SHIFT_LARGE64; r |= (sgn | SPECIAL_ENCODING_MASK64); // add coeff, without leading bits mask = (mask >> 2) - 1; coeff &= mask; r |= coeff; return r; } // // No overflow/underflow checking // __BID_INLINE__ UINT64 fast_get_BID64 (UINT64 sgn, int expon, UINT64 coeff) { UINT64 r, mask; mask = 1; mask <<= EXPONENT_SHIFT_SMALL64; // check whether coefficient fits in 10*5+3 bits if (coeff < mask) { r = expon; r <<= EXPONENT_SHIFT_SMALL64; r |= (coeff | sgn); return r; } // special format // eliminate the case coeff==10^16 after rounding if (coeff == 10000000000000000ull) { r = expon + 1; r <<= EXPONENT_SHIFT_SMALL64; r |= (1000000000000000ull | sgn); return r; } r = expon; r <<= EXPONENT_SHIFT_LARGE64; r |= (sgn | SPECIAL_ENCODING_MASK64); // add coeff, without leading bits mask = (mask >> 2) - 1; coeff &= mask; r |= coeff; return r; } // // no underflow checking // __BID_INLINE__ UINT64 fast_get_BID64_check_OF (UINT64 sgn, int expon, UINT64 coeff, int rmode, unsigned *fpsc) { UINT64 r, mask; if (((unsigned) expon) >= 3 * 256 - 1) { if ((expon == 3 * 256 - 1) && coeff == 10000000000000000ull) { expon = 3 * 256; coeff = 1000000000000000ull; } if (((unsigned) expon) >= 3 * 256) { while (coeff < 1000000000000000ull && expon >= 3 * 256) { expon--; coeff = (coeff << 3) + (coeff << 1); } if (expon > DECIMAL_MAX_EXPON_64) { #ifdef SET_STATUS_FLAGS __set_status_flags (fpsc, OVERFLOW_EXCEPTION | INEXACT_EXCEPTION); #endif // overflow r = sgn | INFINITY_MASK64; switch (rmode) { case ROUNDING_DOWN: if (!sgn) r = LARGEST_BID64; break; case ROUNDING_TO_ZERO: r = sgn | LARGEST_BID64; break; case ROUNDING_UP: // round up if (sgn) r = SMALLEST_BID64; } return r; } } } mask = 1; mask <<= EXPONENT_SHIFT_SMALL64; // check whether coefficient fits in 10*5+3 bits if (coeff < mask) { r = expon; r <<= EXPONENT_SHIFT_SMALL64; r |= (coeff | sgn); return r; } // special format // eliminate the case coeff==10^16 after rounding if (coeff == 10000000000000000ull) { r = expon + 1; r <<= EXPONENT_SHIFT_SMALL64; r |= (1000000000000000ull | sgn); return r; } r = expon; r <<= EXPONENT_SHIFT_LARGE64; r |= (sgn | SPECIAL_ENCODING_MASK64); // add coeff, without leading bits mask = (mask >> 2) - 1; coeff &= mask; r |= coeff; return r; } // // No overflow/underflow checking // or checking for coefficients equal to 10^16 (after rounding) // __BID_INLINE__ UINT64 very_fast_get_BID64 (UINT64 sgn, int expon, UINT64 coeff) { UINT64 r, mask; mask = 1; mask <<= EXPONENT_SHIFT_SMALL64; // check whether coefficient fits in 10*5+3 bits if (coeff < mask) { r = expon; r <<= EXPONENT_SHIFT_SMALL64; r |= (coeff | sgn); return r; } // special format r = expon; r <<= EXPONENT_SHIFT_LARGE64; r |= (sgn | SPECIAL_ENCODING_MASK64); // add coeff, without leading bits mask = (mask >> 2) - 1; coeff &= mask; r |= coeff; return r; } // // No overflow/underflow checking or checking for coefficients above 2^53 // __BID_INLINE__ UINT64 very_fast_get_BID64_small_mantissa (UINT64 sgn, int expon, UINT64 coeff) { // no UF/OF UINT64 r; r = expon; r <<= EXPONENT_SHIFT_SMALL64; r |= (coeff | sgn); return r; } // // This pack macro is used when underflow is known to occur // __BID_INLINE__ UINT64 get_BID64_UF (UINT64 sgn, int expon, UINT64 coeff, UINT64 R, int rmode, unsigned *fpsc) { UINT128 C128, Q_low, Stemp; UINT64 C64, remainder_h, QH, carry, CY; int extra_digits, amount, amount2; unsigned status; // underflow if (expon + MAX_FORMAT_DIGITS < 0) { #ifdef SET_STATUS_FLAGS __set_status_flags (fpsc, UNDERFLOW_EXCEPTION | INEXACT_EXCEPTION); #endif #ifndef IEEE_ROUND_NEAREST_TIES_AWAY #ifndef IEEE_ROUND_NEAREST if (rmode == ROUNDING_DOWN && sgn) return 0x8000000000000001ull; if (rmode == ROUNDING_UP && !sgn) return 1ull; #endif #endif // result is 0 return sgn; } // 10*coeff coeff = (coeff << 3) + (coeff << 1); #ifndef IEEE_ROUND_NEAREST_TIES_AWAY #ifndef IEEE_ROUND_NEAREST if (sgn && (unsigned) (rmode - 1) < 2) rmode = 3 - rmode; #endif #endif if (R) coeff |= 1; // get digits to be shifted out extra_digits = 1 - expon; C128.w[0] = coeff + round_const_table[rmode][extra_digits]; // get coeff*(2^M[extra_digits])/10^extra_digits __mul_64x128_full (QH, Q_low, C128.w[0], reciprocals10_128[extra_digits]); // now get P/10^extra_digits: shift Q_high right by M[extra_digits]-128 amount = recip_scale[extra_digits]; C64 = QH >> amount; //__shr_128(C128, Q_high, amount); #ifndef IEEE_ROUND_NEAREST_TIES_AWAY #ifndef IEEE_ROUND_NEAREST if (rmode == 0) //ROUNDING_TO_NEAREST #endif if (C64 & 1) { // check whether fractional part of initial_P/10^extra_digits is exactly .5 // get remainder amount2 = 64 - amount; remainder_h = 0; remainder_h--; remainder_h >>= amount2; remainder_h = remainder_h & QH; if (!remainder_h && (Q_low.w[1] < reciprocals10_128[extra_digits].w[1] || (Q_low.w[1] == reciprocals10_128[extra_digits].w[1] && Q_low.w[0] < reciprocals10_128[extra_digits].w[0]))) { C64--; } } #endif #ifdef SET_STATUS_FLAGS if (is_inexact (fpsc)) __set_status_flags (fpsc, UNDERFLOW_EXCEPTION); else { status = INEXACT_EXCEPTION; // get remainder remainder_h = QH << (64 - amount); switch (rmode) { case ROUNDING_TO_NEAREST: case ROUNDING_TIES_AWAY: // test whether fractional part is 0 if (remainder_h == 0x8000000000000000ull && (Q_low.w[1] < reciprocals10_128[extra_digits].w[1] || (Q_low.w[1] == reciprocals10_128[extra_digits].w[1] && Q_low.w[0] < reciprocals10_128[extra_digits].w[0]))) status = EXACT_STATUS; break; case ROUNDING_DOWN: case ROUNDING_TO_ZERO: if (!remainder_h && (Q_low.w[1] < reciprocals10_128[extra_digits].w[1] || (Q_low.w[1] == reciprocals10_128[extra_digits].w[1] && Q_low.w[0] < reciprocals10_128[extra_digits].w[0]))) status = EXACT_STATUS; break; default: // round up __add_carry_out (Stemp.w[0], CY, Q_low.w[0], reciprocals10_128[extra_digits].w[0]); __add_carry_in_out (Stemp.w[1], carry, Q_low.w[1], reciprocals10_128[extra_digits].w[1], CY); if ((remainder_h >> (64 - amount)) + carry >= (((UINT64) 1) << amount)) status = EXACT_STATUS; } if (status != EXACT_STATUS) __set_status_flags (fpsc, UNDERFLOW_EXCEPTION | status); } #endif return sgn | C64; } // // This pack macro doesnot check for coefficients above 2^53 // __BID_INLINE__ UINT64 get_BID64_small_mantissa (UINT64 sgn, int expon, UINT64 coeff, int rmode, unsigned *fpsc) { UINT128 C128, Q_low, Stemp; UINT64 r, mask, C64, remainder_h, QH, carry, CY; int extra_digits, amount, amount2; unsigned status; // check for possible underflow/overflow if (((unsigned) expon) >= 3 * 256) { if (expon < 0) { // underflow if (expon + MAX_FORMAT_DIGITS < 0) { #ifdef SET_STATUS_FLAGS __set_status_flags (fpsc, UNDERFLOW_EXCEPTION | INEXACT_EXCEPTION); #endif #ifndef IEEE_ROUND_NEAREST_TIES_AWAY #ifndef IEEE_ROUND_NEAREST if (rmode == ROUNDING_DOWN && sgn) return 0x8000000000000001ull; if (rmode == ROUNDING_UP && !sgn) return 1ull; #endif #endif // result is 0 return sgn; } #ifndef IEEE_ROUND_NEAREST_TIES_AWAY #ifndef IEEE_ROUND_NEAREST if (sgn && (unsigned) (rmode - 1) < 2) rmode = 3 - rmode; #endif #endif // get digits to be shifted out extra_digits = -expon; C128.w[0] = coeff + round_const_table[rmode][extra_digits]; // get coeff*(2^M[extra_digits])/10^extra_digits __mul_64x128_full (QH, Q_low, C128.w[0], reciprocals10_128[extra_digits]); // now get P/10^extra_digits: shift Q_high right by M[extra_digits]-128 amount = recip_scale[extra_digits]; C64 = QH >> amount; #ifndef IEEE_ROUND_NEAREST_TIES_AWAY #ifndef IEEE_ROUND_NEAREST if (rmode == 0) //ROUNDING_TO_NEAREST #endif if (C64 & 1) { // check whether fractional part of initial_P/10^extra_digits is exactly .5 // get remainder amount2 = 64 - amount; remainder_h = 0; remainder_h--; remainder_h >>= amount2; remainder_h = remainder_h & QH; if (!remainder_h && (Q_low.w[1] < reciprocals10_128[extra_digits].w[1] || (Q_low.w[1] == reciprocals10_128[extra_digits].w[1] && Q_low.w[0] < reciprocals10_128[extra_digits].w[0]))) { C64--; } } #endif #ifdef SET_STATUS_FLAGS if (is_inexact (fpsc)) __set_status_flags (fpsc, UNDERFLOW_EXCEPTION); else { status = INEXACT_EXCEPTION; // get remainder remainder_h = QH << (64 - amount); switch (rmode) { case ROUNDING_TO_NEAREST: case ROUNDING_TIES_AWAY: // test whether fractional part is 0 if (remainder_h == 0x8000000000000000ull && (Q_low.w[1] < reciprocals10_128[extra_digits].w[1] || (Q_low.w[1] == reciprocals10_128[extra_digits].w[1] && Q_low.w[0] < reciprocals10_128[extra_digits].w[0]))) status = EXACT_STATUS; break; case ROUNDING_DOWN: case ROUNDING_TO_ZERO: if (!remainder_h && (Q_low.w[1] < reciprocals10_128[extra_digits].w[1] || (Q_low.w[1] == reciprocals10_128[extra_digits].w[1] && Q_low.w[0] < reciprocals10_128[extra_digits].w[0]))) status = EXACT_STATUS; break; default: // round up __add_carry_out (Stemp.w[0], CY, Q_low.w[0], reciprocals10_128[extra_digits].w[0]); __add_carry_in_out (Stemp.w[1], carry, Q_low.w[1], reciprocals10_128[extra_digits].w[1], CY); if ((remainder_h >> (64 - amount)) + carry >= (((UINT64) 1) << amount)) status = EXACT_STATUS; } if (status != EXACT_STATUS) __set_status_flags (fpsc, UNDERFLOW_EXCEPTION | status); } #endif return sgn | C64; } while (coeff < 1000000000000000ull && expon >= 3 * 256) { expon--; coeff = (coeff << 3) + (coeff << 1); } if (expon > DECIMAL_MAX_EXPON_64) { #ifdef SET_STATUS_FLAGS __set_status_flags (fpsc, OVERFLOW_EXCEPTION | INEXACT_EXCEPTION); #endif // overflow r = sgn | INFINITY_MASK64; switch (rmode) { case ROUNDING_DOWN: if (!sgn) r = LARGEST_BID64; break; case ROUNDING_TO_ZERO: r = sgn | LARGEST_BID64; break; case ROUNDING_UP: // round up if (sgn) r = SMALLEST_BID64; } return r; } else { mask = 1; mask <<= EXPONENT_SHIFT_SMALL64; if (coeff >= mask) { r = expon; r <<= EXPONENT_SHIFT_LARGE64; r |= (sgn | SPECIAL_ENCODING_MASK64); // add coeff, without leading bits mask = (mask >> 2) - 1; coeff &= mask; r |= coeff; return r; } } } r = expon; r <<= EXPONENT_SHIFT_SMALL64; r |= (coeff | sgn); return r; } /***************************************************************************** * * BID128 pack/unpack macros * *****************************************************************************/ // // Macro for handling BID128 underflow // sticky bit given as additional argument // __BID_INLINE__ UINT128 * handle_UF_128_rem (UINT128 * pres, UINT64 sgn, int expon, UINT128 CQ, UINT64 R, unsigned *prounding_mode, unsigned *fpsc) { UINT128 T128, TP128, Qh, Ql, Qh1, Stemp, Tmp, Tmp1, CQ2, CQ8; UINT64 carry, CY; int ed2, amount; unsigned rmode, status; // UF occurs if (expon + MAX_FORMAT_DIGITS_128 < 0) { #ifdef SET_STATUS_FLAGS __set_status_flags (fpsc, UNDERFLOW_EXCEPTION | INEXACT_EXCEPTION); #endif pres->w[1] = sgn; pres->w[0] = 0; #ifndef IEEE_ROUND_NEAREST_TIES_AWAY #ifndef IEEE_ROUND_NEAREST if ((sgn && *prounding_mode == ROUNDING_DOWN) || (!sgn && *prounding_mode == ROUNDING_UP)) pres->w[0] = 1ull; #endif #endif return pres; } // CQ *= 10 CQ2.w[1] = (CQ.w[1] << 1) | (CQ.w[0] >> 63); CQ2.w[0] = CQ.w[0] << 1; CQ8.w[1] = (CQ.w[1] << 3) | (CQ.w[0] >> 61); CQ8.w[0] = CQ.w[0] << 3; __add_128_128 (CQ, CQ2, CQ8); // add remainder if (R) CQ.w[0] |= 1; ed2 = 1 - expon; // add rounding constant to CQ #ifndef IEEE_ROUND_NEAREST_TIES_AWAY #ifndef IEEE_ROUND_NEAREST rmode = *prounding_mode; if (sgn && (unsigned) (rmode - 1) < 2) rmode = 3 - rmode; #else rmode = 0; #endif #else rmode = 0; #endif T128 = round_const_table_128[rmode][ed2]; __add_carry_out (CQ.w[0], carry, T128.w[0], CQ.w[0]); CQ.w[1] = CQ.w[1] + T128.w[1] + carry; TP128 = reciprocals10_128[ed2]; __mul_128x128_full (Qh, Ql, CQ, TP128); amount = recip_scale[ed2]; if (amount >= 64) { CQ.w[0] = Qh.w[1] >> (amount - 64); CQ.w[1] = 0; } else { __shr_128 (CQ, Qh, amount); } expon = 0; #ifndef IEEE_ROUND_NEAREST_TIES_AWAY #ifndef IEEE_ROUND_NEAREST if (!(*prounding_mode)) #endif if (CQ.w[0] & 1) { // check whether fractional part of initial_P/10^ed1 is exactly .5 // get remainder __shl_128_long (Qh1, Qh, (128 - amount)); if (!Qh1.w[1] && !Qh1.w[0] && (Ql.w[1] < reciprocals10_128[ed2].w[1] || (Ql.w[1] == reciprocals10_128[ed2].w[1] && Ql.w[0] < reciprocals10_128[ed2].w[0]))) { CQ.w[0]--; } } #endif #ifdef SET_STATUS_FLAGS if (is_inexact (fpsc)) __set_status_flags (fpsc, UNDERFLOW_EXCEPTION); else { status = INEXACT_EXCEPTION; // get remainder __shl_128_long (Qh1, Qh, (128 - amount)); switch (rmode) { case ROUNDING_TO_NEAREST: case ROUNDING_TIES_AWAY: // test whether fractional part is 0 if (Qh1.w[1] == 0x8000000000000000ull && (!Qh1.w[0]) && (Ql.w[1] < reciprocals10_128[ed2].w[1] || (Ql.w[1] == reciprocals10_128[ed2].w[1] && Ql.w[0] < reciprocals10_128[ed2].w[0]))) status = EXACT_STATUS; break; case ROUNDING_DOWN: case ROUNDING_TO_ZERO: if ((!Qh1.w[1]) && (!Qh1.w[0]) && (Ql.w[1] < reciprocals10_128[ed2].w[1] || (Ql.w[1] == reciprocals10_128[ed2].w[1] && Ql.w[0] < reciprocals10_128[ed2].w[0]))) status = EXACT_STATUS; break; default: // round up __add_carry_out (Stemp.w[0], CY, Ql.w[0], reciprocals10_128[ed2].w[0]); __add_carry_in_out (Stemp.w[1], carry, Ql.w[1], reciprocals10_128[ed2].w[1], CY); __shr_128_long (Qh, Qh1, (128 - amount)); Tmp.w[0] = 1; Tmp.w[1] = 0; __shl_128_long (Tmp1, Tmp, amount); Qh.w[0] += carry; if (Qh.w[0] < carry) Qh.w[1]++; if (__unsigned_compare_ge_128 (Qh, Tmp1)) status = EXACT_STATUS; } if (status != EXACT_STATUS) __set_status_flags (fpsc, UNDERFLOW_EXCEPTION | status); } #endif pres->w[1] = sgn | CQ.w[1]; pres->w[0] = CQ.w[0]; return pres; } // // Macro for handling BID128 underflow // __BID_INLINE__ UINT128 * handle_UF_128 (UINT128 * pres, UINT64 sgn, int expon, UINT128 CQ, unsigned *prounding_mode, unsigned *fpsc) { UINT128 T128, TP128, Qh, Ql, Qh1, Stemp, Tmp, Tmp1; UINT64 carry, CY; int ed2, amount; unsigned rmode, status; // UF occurs if (expon + MAX_FORMAT_DIGITS_128 < 0) { #ifdef SET_STATUS_FLAGS __set_status_flags (fpsc, UNDERFLOW_EXCEPTION | INEXACT_EXCEPTION); #endif pres->w[1] = sgn; pres->w[0] = 0; #ifndef IEEE_ROUND_NEAREST_TIES_AWAY #ifndef IEEE_ROUND_NEAREST if ((sgn && *prounding_mode == ROUNDING_DOWN) || (!sgn && *prounding_mode == ROUNDING_UP)) pres->w[0] = 1ull; #endif #endif return pres; } ed2 = 0 - expon; // add rounding constant to CQ #ifndef IEEE_ROUND_NEAREST_TIES_AWAY #ifndef IEEE_ROUND_NEAREST rmode = *prounding_mode; if (sgn && (unsigned) (rmode - 1) < 2) rmode = 3 - rmode; #else rmode = 0; #endif #else rmode = 0; #endif T128 = round_const_table_128[rmode][ed2]; __add_carry_out (CQ.w[0], carry, T128.w[0], CQ.w[0]); CQ.w[1] = CQ.w[1] + T128.w[1] + carry; TP128 = reciprocals10_128[ed2]; __mul_128x128_full (Qh, Ql, CQ, TP128); amount = recip_scale[ed2]; if (amount >= 64) { CQ.w[0] = Qh.w[1] >> (amount - 64); CQ.w[1] = 0; } else { __shr_128 (CQ, Qh, amount); } expon = 0; #ifndef IEEE_ROUND_NEAREST_TIES_AWAY #ifndef IEEE_ROUND_NEAREST if (!(*prounding_mode)) #endif if (CQ.w[0] & 1) { // check whether fractional part of initial_P/10^ed1 is exactly .5 // get remainder __shl_128_long (Qh1, Qh, (128 - amount)); if (!Qh1.w[1] && !Qh1.w[0] && (Ql.w[1] < reciprocals10_128[ed2].w[1] || (Ql.w[1] == reciprocals10_128[ed2].w[1] && Ql.w[0] < reciprocals10_128[ed2].w[0]))) { CQ.w[0]--; } } #endif #ifdef SET_STATUS_FLAGS if (is_inexact (fpsc)) __set_status_flags (fpsc, UNDERFLOW_EXCEPTION); else { status = INEXACT_EXCEPTION; // get remainder __shl_128_long (Qh1, Qh, (128 - amount)); switch (rmode) { case ROUNDING_TO_NEAREST: case ROUNDING_TIES_AWAY: // test whether fractional part is 0 if (Qh1.w[1] == 0x8000000000000000ull && (!Qh1.w[0]) && (Ql.w[1] < reciprocals10_128[ed2].w[1] || (Ql.w[1] == reciprocals10_128[ed2].w[1] && Ql.w[0] < reciprocals10_128[ed2].w[0]))) status = EXACT_STATUS; break; case ROUNDING_DOWN: case ROUNDING_TO_ZERO: if ((!Qh1.w[1]) && (!Qh1.w[0]) && (Ql.w[1] < reciprocals10_128[ed2].w[1] || (Ql.w[1] == reciprocals10_128[ed2].w[1] && Ql.w[0] < reciprocals10_128[ed2].w[0]))) status = EXACT_STATUS; break; default: // round up __add_carry_out (Stemp.w[0], CY, Ql.w[0], reciprocals10_128[ed2].w[0]); __add_carry_in_out (Stemp.w[1], carry, Ql.w[1], reciprocals10_128[ed2].w[1], CY); __shr_128_long (Qh, Qh1, (128 - amount)); Tmp.w[0] = 1; Tmp.w[1] = 0; __shl_128_long (Tmp1, Tmp, amount); Qh.w[0] += carry; if (Qh.w[0] < carry) Qh.w[1]++; if (__unsigned_compare_ge_128 (Qh, Tmp1)) status = EXACT_STATUS; } if (status != EXACT_STATUS) __set_status_flags (fpsc, UNDERFLOW_EXCEPTION | status); } #endif pres->w[1] = sgn | CQ.w[1]; pres->w[0] = CQ.w[0]; return pres; } // // BID128 unpack, input passed by value // __BID_INLINE__ UINT64 unpack_BID128_value (UINT64 * psign_x, int *pexponent_x, UINT128 * pcoefficient_x, UINT128 x) { UINT128 coeff, T33, T34; UINT64 ex; *psign_x = (x.w[1]) & 0x8000000000000000ull; // special encodings if ((x.w[1] & INFINITY_MASK64) >= SPECIAL_ENCODING_MASK64) { if ((x.w[1] & INFINITY_MASK64) < INFINITY_MASK64) { // non-canonical input pcoefficient_x->w[0] = 0; pcoefficient_x->w[1] = 0; ex = (x.w[1]) >> 47; *pexponent_x = ((int) ex) & EXPONENT_MASK128; return 0; } // 10^33 T33 = power10_table_128[33]; /*coeff.w[0] = x.w[0]; coeff.w[1] = (x.w[1]) & LARGE_COEFF_MASK128; pcoefficient_x->w[0] = x.w[0]; pcoefficient_x->w[1] = x.w[1]; if (__unsigned_compare_ge_128 (coeff, T33)) // non-canonical pcoefficient_x->w[1] &= (~LARGE_COEFF_MASK128); */ pcoefficient_x->w[0] = x.w[0]; pcoefficient_x->w[1] = (x.w[1]) & 0x00003fffffffffffull; if (__unsigned_compare_ge_128 ((*pcoefficient_x), T33)) // non-canonical { pcoefficient_x->w[1] = (x.w[1]) & 0xfe00000000000000ull; pcoefficient_x->w[0] = 0; } else pcoefficient_x->w[1] = (x.w[1]) & 0xfe003fffffffffffull; if ((x.w[1] & NAN_MASK64) == INFINITY_MASK64) { pcoefficient_x->w[0] = 0; pcoefficient_x->w[1] = x.w[1] & SINFINITY_MASK64; } *pexponent_x = 0; return 0; // NaN or Infinity } coeff.w[0] = x.w[0]; coeff.w[1] = (x.w[1]) & SMALL_COEFF_MASK128; // 10^34 T34 = power10_table_128[34]; // check for non-canonical values if (__unsigned_compare_ge_128 (coeff, T34)) coeff.w[0] = coeff.w[1] = 0; pcoefficient_x->w[0] = coeff.w[0]; pcoefficient_x->w[1] = coeff.w[1]; ex = (x.w[1]) >> 49; *pexponent_x = ((int) ex) & EXPONENT_MASK128; return coeff.w[0] | coeff.w[1]; } // // BID128 unpack, input pased by reference // __BID_INLINE__ UINT64 unpack_BID128 (UINT64 * psign_x, int *pexponent_x, UINT128 * pcoefficient_x, UINT128 * px) { UINT128 coeff, T33, T34; UINT64 ex; *psign_x = (px->w[1]) & 0x8000000000000000ull; // special encodings if ((px->w[1] & INFINITY_MASK64) >= SPECIAL_ENCODING_MASK64) { if ((px->w[1] & INFINITY_MASK64) < INFINITY_MASK64) { // non-canonical input pcoefficient_x->w[0] = 0; pcoefficient_x->w[1] = 0; ex = (px->w[1]) >> 47; *pexponent_x = ((int) ex) & EXPONENT_MASK128; return 0; } // 10^33 T33 = power10_table_128[33]; coeff.w[0] = px->w[0]; coeff.w[1] = (px->w[1]) & LARGE_COEFF_MASK128; pcoefficient_x->w[0] = px->w[0]; pcoefficient_x->w[1] = px->w[1]; if (__unsigned_compare_ge_128 (coeff, T33)) { // non-canonical pcoefficient_x->w[1] &= (~LARGE_COEFF_MASK128); pcoefficient_x->w[0] = 0; } *pexponent_x = 0; return 0; // NaN or Infinity } coeff.w[0] = px->w[0]; coeff.w[1] = (px->w[1]) & SMALL_COEFF_MASK128; // 10^34 T34 = power10_table_128[34]; // check for non-canonical values if (__unsigned_compare_ge_128 (coeff, T34)) coeff.w[0] = coeff.w[1] = 0; pcoefficient_x->w[0] = coeff.w[0]; pcoefficient_x->w[1] = coeff.w[1]; ex = (px->w[1]) >> 49; *pexponent_x = ((int) ex) & EXPONENT_MASK128; return coeff.w[0] | coeff.w[1]; } // // Pack macro checks for overflow, but not underflow // __BID_INLINE__ UINT128 * get_BID128_very_fast_OF (UINT128 * pres, UINT64 sgn, int expon, UINT128 coeff, unsigned *prounding_mode, unsigned *fpsc) { UINT128 T; UINT64 tmp, tmp2; if ((unsigned) expon > DECIMAL_MAX_EXPON_128) { if (expon - MAX_FORMAT_DIGITS_128 <= DECIMAL_MAX_EXPON_128) { T = power10_table_128[MAX_FORMAT_DIGITS_128 - 1]; while (__unsigned_compare_gt_128 (T, coeff) && expon > DECIMAL_MAX_EXPON_128) { coeff.w[1] = (coeff.w[1] << 3) + (coeff.w[1] << 1) + (coeff.w[0] >> 61) + (coeff.w[0] >> 63); tmp2 = coeff.w[0] << 3; coeff.w[0] = (coeff.w[0] << 1) + tmp2; if (coeff.w[0] < tmp2) coeff.w[1]++; expon--; } } if ((unsigned) expon > DECIMAL_MAX_EXPON_128) { // OF #ifdef SET_STATUS_FLAGS __set_status_flags (fpsc, OVERFLOW_EXCEPTION | INEXACT_EXCEPTION); #endif #ifndef IEEE_ROUND_NEAREST_TIES_AWAY #ifndef IEEE_ROUND_NEAREST if (*prounding_mode == ROUNDING_TO_ZERO || (sgn && *prounding_mode == ROUNDING_UP) || (!sgn && *prounding_mode == ROUNDING_DOWN)) { pres->w[1] = sgn | LARGEST_BID128_HIGH; pres->w[0] = LARGEST_BID128_LOW; } else #endif #endif { pres->w[1] = sgn | INFINITY_MASK64; pres->w[0] = 0; } return pres; } } pres->w[0] = coeff.w[0]; tmp = expon; tmp <<= 49; pres->w[1] = sgn | tmp | coeff.w[1]; return pres; } // // No overflow/underflow checks // No checking for coefficient == 10^34 (rounding artifact) // __BID_INLINE__ UINT128 * get_BID128_very_fast (UINT128 * pres, UINT64 sgn, int expon, UINT128 coeff) { UINT64 tmp; pres->w[0] = coeff.w[0]; tmp = expon; tmp <<= 49; pres->w[1] = sgn | tmp | coeff.w[1]; return pres; } // // No overflow/underflow checks // __BID_INLINE__ UINT128 * get_BID128_fast (UINT128 * pres, UINT64 sgn, int expon, UINT128 coeff) { UINT64 tmp; // coeff==10^34? if (coeff.w[1] == 0x0001ed09bead87c0ull && coeff.w[0] == 0x378d8e6400000000ull) { expon++; // set coefficient to 10^33 coeff.w[1] = 0x0000314dc6448d93ull; coeff.w[0] = 0x38c15b0a00000000ull; } pres->w[0] = coeff.w[0]; tmp = expon; tmp <<= 49; pres->w[1] = sgn | tmp | coeff.w[1]; return pres; } // // General BID128 pack macro // __BID_INLINE__ UINT128 * get_BID128 (UINT128 * pres, UINT64 sgn, int expon, UINT128 coeff, unsigned *prounding_mode, unsigned *fpsc) { UINT128 T; UINT64 tmp, tmp2; // coeff==10^34? if (coeff.w[1] == 0x0001ed09bead87c0ull && coeff.w[0] == 0x378d8e6400000000ull) { expon++; // set coefficient to 10^33 coeff.w[1] = 0x0000314dc6448d93ull; coeff.w[0] = 0x38c15b0a00000000ull; } // check OF, UF if (expon < 0 || expon > DECIMAL_MAX_EXPON_128) { // check UF if (expon < 0) { return handle_UF_128 (pres, sgn, expon, coeff, prounding_mode, fpsc); } if (expon - MAX_FORMAT_DIGITS_128 <= DECIMAL_MAX_EXPON_128) { T = power10_table_128[MAX_FORMAT_DIGITS_128 - 1]; while (__unsigned_compare_gt_128 (T, coeff) && expon > DECIMAL_MAX_EXPON_128) { coeff.w[1] = (coeff.w[1] << 3) + (coeff.w[1] << 1) + (coeff.w[0] >> 61) + (coeff.w[0] >> 63); tmp2 = coeff.w[0] << 3; coeff.w[0] = (coeff.w[0] << 1) + tmp2; if (coeff.w[0] < tmp2) coeff.w[1]++; expon--; } } if (expon > DECIMAL_MAX_EXPON_128) { if (!(coeff.w[1] | coeff.w[0])) { pres->w[1] = sgn | (((UINT64) DECIMAL_MAX_EXPON_128) << 49); pres->w[0] = 0; return pres; } // OF #ifdef SET_STATUS_FLAGS __set_status_flags (fpsc, OVERFLOW_EXCEPTION | INEXACT_EXCEPTION); #endif #ifndef IEEE_ROUND_NEAREST_TIES_AWAY #ifndef IEEE_ROUND_NEAREST if (*prounding_mode == ROUNDING_TO_ZERO || (sgn && *prounding_mode == ROUNDING_UP) || (!sgn && *prounding_mode == ROUNDING_DOWN)) { pres->w[1] = sgn | LARGEST_BID128_HIGH; pres->w[0] = LARGEST_BID128_LOW; } else #endif #endif { pres->w[1] = sgn | INFINITY_MASK64; pres->w[0] = 0; } return pres; } } pres->w[0] = coeff.w[0]; tmp = expon; tmp <<= 49; pres->w[1] = sgn | tmp | coeff.w[1]; return pres; } // // Macro used for conversions from string // (no additional arguments given for rounding mode, status flags) // __BID_INLINE__ UINT128 * get_BID128_string (UINT128 * pres, UINT64 sgn, int expon, UINT128 coeff) { UINT128 D2, D8; UINT64 tmp; unsigned rmode = 0, status; // coeff==10^34? if (coeff.w[1] == 0x0001ed09bead87c0ull && coeff.w[0] == 0x378d8e6400000000ull) { expon++; // set coefficient to 10^33 coeff.w[1] = 0x0000314dc6448d93ull; coeff.w[0] = 0x38c15b0a00000000ull; } // check OF, UF if ((unsigned) expon > DECIMAL_MAX_EXPON_128) { // check UF if (expon < 0) return handle_UF_128 (pres, sgn, expon, coeff, &rmode, &status); // OF if (expon < DECIMAL_MAX_EXPON_128 + 34) { while (expon > DECIMAL_MAX_EXPON_128 && (coeff.w[1] < power10_table_128[33].w[1] || (coeff.w[1] == power10_table_128[33].w[1] && coeff.w[0] < power10_table_128[33].w[0]))) { D2.w[1] = (coeff.w[1] << 1) | (coeff.w[0] >> 63); D2.w[0] = coeff.w[0] << 1; D8.w[1] = (coeff.w[1] << 3) | (coeff.w[0] >> 61); D8.w[0] = coeff.w[0] << 3; __add_128_128 (coeff, D2, D8); expon--; } } else if (!(coeff.w[0] | coeff.w[1])) expon = DECIMAL_MAX_EXPON_128; if (expon > DECIMAL_MAX_EXPON_128) { pres->w[1] = sgn | INFINITY_MASK64; pres->w[0] = 0; switch (rmode) { case ROUNDING_DOWN: if (!sgn) { pres->w[1] = LARGEST_BID128_HIGH; pres->w[0] = LARGEST_BID128_LOW; } break; case ROUNDING_TO_ZERO: pres->w[1] = sgn | LARGEST_BID128_HIGH; pres->w[0] = LARGEST_BID128_LOW; break; case ROUNDING_UP: // round up if (sgn) { pres->w[1] = sgn | LARGEST_BID128_HIGH; pres->w[0] = LARGEST_BID128_LOW; } break; } return pres; } } pres->w[0] = coeff.w[0]; tmp = expon; tmp <<= 49; pres->w[1] = sgn | tmp | coeff.w[1]; return pres; } /***************************************************************************** * * BID32 pack/unpack macros * *****************************************************************************/ __BID_INLINE__ UINT32 unpack_BID32 (UINT32 * psign_x, int *pexponent_x, UINT32 * pcoefficient_x, UINT32 x) { UINT32 tmp; *psign_x = x & 0x80000000; if ((x & SPECIAL_ENCODING_MASK32) == SPECIAL_ENCODING_MASK32) { // special encodings if ((x & INFINITY_MASK32) == INFINITY_MASK32) { *pcoefficient_x = x & 0xfe0fffff; if ((x & 0x000fffff) >= 1000000) *pcoefficient_x = x & 0xfe000000; if ((x & NAN_MASK32) == INFINITY_MASK32) *pcoefficient_x = x & 0xf8000000; *pexponent_x = 0; return 0; // NaN or Infinity } // coefficient *pcoefficient_x = (x & SMALL_COEFF_MASK32) | LARGE_COEFF_HIGH_BIT32; // check for non-canonical value if (*pcoefficient_x >= 10000000) *pcoefficient_x = 0; // get exponent tmp = x >> 21; *pexponent_x = tmp & EXPONENT_MASK32; return 1; } // exponent tmp = x >> 23; *pexponent_x = tmp & EXPONENT_MASK32; // coefficient *pcoefficient_x = (x & LARGE_COEFF_MASK32); return *pcoefficient_x; } // // General pack macro for BID32 // __BID_INLINE__ UINT32 get_BID32 (UINT32 sgn, int expon, UINT64 coeff, int rmode, unsigned *fpsc) { UINT128 Q; UINT64 C64, remainder_h, carry, Stemp; UINT32 r, mask; int extra_digits, amount, amount2; unsigned status; if (coeff > 9999999ull) { expon++; coeff = 1000000ull; } // check for possible underflow/overflow if (((unsigned) expon) > DECIMAL_MAX_EXPON_32) { if (expon < 0) { // underflow if (expon + MAX_FORMAT_DIGITS_32 < 0) { #ifdef SET_STATUS_FLAGS __set_status_flags (fpsc, UNDERFLOW_EXCEPTION | INEXACT_EXCEPTION); #endif #ifndef IEEE_ROUND_NEAREST_TIES_AWAY #ifndef IEEE_ROUND_NEAREST if (rmode == ROUNDING_DOWN && sgn) return 0x80000001; if (rmode == ROUNDING_UP && !sgn) return 1; #endif #endif // result is 0 return sgn; } // get digits to be shifted out #ifdef IEEE_ROUND_NEAREST_TIES_AWAY rmode = 0; #endif #ifdef IEEE_ROUND_NEAREST rmode = 0; #endif #ifndef IEEE_ROUND_NEAREST_TIES_AWAY #ifndef IEEE_ROUND_NEAREST if (sgn && (unsigned) (rmode - 1) < 2) rmode = 3 - rmode; #endif #endif extra_digits = -expon; coeff += round_const_table[rmode][extra_digits]; // get coeff*(2^M[extra_digits])/10^extra_digits __mul_64x64_to_128 (Q, coeff, reciprocals10_64[extra_digits]); // now get P/10^extra_digits: shift Q_high right by M[extra_digits]-128 amount = short_recip_scale[extra_digits]; C64 = Q.w[1] >> amount; #ifndef IEEE_ROUND_NEAREST_TIES_AWAY #ifndef IEEE_ROUND_NEAREST if (rmode == 0) //ROUNDING_TO_NEAREST #endif if (C64 & 1) { // check whether fractional part of initial_P/10^extra_digits is exactly .5 // get remainder amount2 = 64 - amount; remainder_h = 0; remainder_h--; remainder_h >>= amount2; remainder_h = remainder_h & Q.w[1]; if (!remainder_h && (Q.w[0] < reciprocals10_64[extra_digits])) { C64--; } } #endif #ifdef SET_STATUS_FLAGS if (is_inexact (fpsc)) __set_status_flags (fpsc, UNDERFLOW_EXCEPTION); else { status = INEXACT_EXCEPTION; // get remainder remainder_h = Q.w[1] << (64 - amount); switch (rmode) { case ROUNDING_TO_NEAREST: case ROUNDING_TIES_AWAY: // test whether fractional part is 0 if (remainder_h == 0x8000000000000000ull && (Q.w[0] < reciprocals10_64[extra_digits])) status = EXACT_STATUS; break; case ROUNDING_DOWN: case ROUNDING_TO_ZERO: if (!remainder_h && (Q.w[0] < reciprocals10_64[extra_digits])) status = EXACT_STATUS; break; default: // round up __add_carry_out (Stemp, carry, Q.w[0], reciprocals10_64[extra_digits]); if ((remainder_h >> (64 - amount)) + carry >= (((UINT64) 1) << amount)) status = EXACT_STATUS; } if (status != EXACT_STATUS) __set_status_flags (fpsc, UNDERFLOW_EXCEPTION | status); } #endif return sgn | (UINT32) C64; } while (coeff < 1000000 && expon > DECIMAL_MAX_EXPON_32) { coeff = (coeff << 3) + (coeff << 1); expon--; } if (((unsigned) expon) > DECIMAL_MAX_EXPON_32) { __set_status_flags (fpsc, OVERFLOW_EXCEPTION | INEXACT_EXCEPTION); // overflow r = sgn | INFINITY_MASK32; switch (rmode) { case ROUNDING_DOWN: if (!sgn) r = LARGEST_BID32; break; case ROUNDING_TO_ZERO: r = sgn | LARGEST_BID32; break; case ROUNDING_UP: // round up if (sgn) r = sgn | LARGEST_BID32; } return r; } } mask = 1 << 23; // check whether coefficient fits in DECIMAL_COEFF_FIT bits if (coeff < mask) { r = expon; r <<= 23; r |= ((UINT32) coeff | sgn); return r; } // special format r = expon; r <<= 21; r |= (sgn | SPECIAL_ENCODING_MASK32); // add coeff, without leading bits mask = (1 << 21) - 1; r |= (((UINT32) coeff) & mask); return r; } // // no overflow/underflow checks // __BID_INLINE__ UINT32 very_fast_get_BID32 (UINT32 sgn, int expon, UINT32 coeff) { UINT32 r, mask; mask = 1 << 23; // check whether coefficient fits in 10*2+3 bits if (coeff < mask) { r = expon; r <<= 23; r |= (coeff | sgn); return r; } // special format r = expon; r <<= 21; r |= (sgn | SPECIAL_ENCODING_MASK32); // add coeff, without leading bits mask = (1 << 21) - 1; coeff &= mask; r |= coeff; return r; } /************************************************************* * *************************************************************/ typedef ALIGN (16) struct { UINT64 w[6]; } UINT384; typedef ALIGN (16) struct { UINT64 w[8]; } UINT512; // #define P 34 #define MASK_STEERING_BITS 0x6000000000000000ull #define MASK_BINARY_EXPONENT1 0x7fe0000000000000ull #define MASK_BINARY_SIG1 0x001fffffffffffffull #define MASK_BINARY_EXPONENT2 0x1ff8000000000000ull //used to take G[2:w+3] (sec 3.3) #define MASK_BINARY_SIG2 0x0007ffffffffffffull //used to mask out G4:T0 (sec 3.3) #define MASK_BINARY_OR2 0x0020000000000000ull //used to prefix 8+G4 to T (sec 3.3) #define UPPER_EXPON_LIMIT 51 #define MASK_EXP 0x7ffe000000000000ull #define MASK_SPECIAL 0x7800000000000000ull #define MASK_NAN 0x7c00000000000000ull #define MASK_SNAN 0x7e00000000000000ull #define MASK_ANY_INF 0x7c00000000000000ull #define MASK_INF 0x7800000000000000ull #define MASK_SIGN 0x8000000000000000ull #define MASK_COEFF 0x0001ffffffffffffull #define BIN_EXP_BIAS (0x1820ull << 49) #define EXP_MIN 0x0000000000000000ull // EXP_MIN = (-6176 + 6176) << 49 #define EXP_MAX 0x5ffe000000000000ull // EXP_MAX = (6111 + 6176) << 49 #define EXP_MAX_P1 0x6000000000000000ull // EXP_MAX + 1 = (6111 + 6176 + 1) << 49 #define EXP_P1 0x0002000000000000ull // EXP_ P1= 1 << 49 #define expmin -6176 // min unbiased exponent #define expmax 6111 // max unbiased exponent #define expmin16 -398 // min unbiased exponent #define expmax16 369 // max unbiased exponent #define SIGNMASK32 0x80000000 #define BID64_SIG_MAX 0x002386F26FC0ffffull #define SIGNMASK64 0x8000000000000000ull // typedef unsigned int FPSC; // floating-point status and control // bit31: // bit30: // bit29: // bit28: // bit27: // bit26: // bit25: // bit24: // bit23: // bit22: // bit21: // bit20: // bit19: // bit18: // bit17: // bit16: // bit15: // bit14: RC:2 // bit13: RC:1 // bit12: RC:0 // bit11: PM // bit10: UM // bit9: OM // bit8: ZM // bit7: DM // bit6: IM // bit5: PE // bit4: UE // bit3: OE // bit2: ZE // bit1: DE // bit0: IE #define ROUNDING_MODE_MASK 0x00007000 typedef struct _DEC_DIGITS { unsigned int digits; UINT64 threshold_hi; UINT64 threshold_lo; unsigned int digits1; } DEC_DIGITS; extern DEC_DIGITS nr_digits[]; extern UINT64 midpoint64[]; extern UINT128 midpoint128[]; extern UINT192 midpoint192[]; extern UINT256 midpoint256[]; extern UINT64 ten2k64[]; extern UINT128 ten2k128[]; extern UINT256 ten2k256[]; extern UINT128 ten2mk128[]; extern UINT64 ten2mk64[]; extern UINT128 ten2mk128trunc[]; extern int shiftright128[]; extern UINT64 maskhigh128[]; extern UINT64 maskhigh128M[]; extern UINT64 maskhigh192M[]; extern UINT64 maskhigh256M[]; extern UINT64 onehalf128[]; extern UINT64 onehalf128M[]; extern UINT64 onehalf192M[]; extern UINT64 onehalf256M[]; extern UINT128 ten2mk128M[]; extern UINT128 ten2mk128truncM[]; extern UINT192 ten2mk192truncM[]; extern UINT256 ten2mk256truncM[]; extern int shiftright128M[]; extern int shiftright192M[]; extern int shiftright256M[]; extern UINT192 ten2mk192M[]; extern UINT256 ten2mk256M[]; extern unsigned char char_table2[]; extern unsigned char char_table3[]; extern UINT64 ten2m3k64[]; extern unsigned int shift_ten2m3k64[]; extern UINT128 ten2m3k128[]; extern unsigned int shift_ten2m3k128[]; /*************************************************************************** *************** TABLES FOR GENERAL ROUNDING FUNCTIONS ********************* ***************************************************************************/ extern UINT64 Kx64[]; extern unsigned int Ex64m64[]; extern UINT64 half64[]; extern UINT64 mask64[]; extern UINT64 ten2mxtrunc64[]; extern UINT128 Kx128[]; extern unsigned int Ex128m128[]; extern UINT64 half128[]; extern UINT64 mask128[]; extern UINT128 ten2mxtrunc128[]; extern UINT192 Kx192[]; extern unsigned int Ex192m192[]; extern UINT64 half192[]; extern UINT64 mask192[]; extern UINT192 ten2mxtrunc192[]; extern UINT256 Kx256[]; extern unsigned int Ex256m256[]; extern UINT64 half256[]; extern UINT64 mask256[]; extern UINT256 ten2mxtrunc256[]; typedef union __bid64_128 { UINT64 b64; UINT128 b128; } BID64_128; BID64_128 bid_fma (unsigned int P0, BID64_128 x1, unsigned int P1, BID64_128 y1, unsigned int P2, BID64_128 z1, unsigned int P3, unsigned int rnd_mode, FPSC * fpsc); #define P16 16 #define P34 34 union __int_double { UINT64 i; double d; }; typedef union __int_double int_double; union __int_float { UINT32 i; float d; }; typedef union __int_float int_float; #define SWAP(A,B,T) {\ T = A; \ A = B; \ B = T; \ } // this macro will find coefficient_x to be in [2^A, 2^(A+1) ) // ie it knows that it is A bits long #define NUMBITS(A, coefficient_x, tempx){\ temp_x.d=(float)coefficient_x;\ A=((tempx.i >>23) & EXPONENT_MASK32) - 0x7f;\ } enum class_types { signalingNaN, quietNaN, negativeInfinity, negativeNormal, negativeSubnormal, negativeZero, positiveZero, positiveSubnormal, positiveNormal, positiveInfinity }; typedef union { UINT64 ui64; double d; } BID_UI64DOUBLE; #endif
Go to most recent revision | Compare with Previous | Blame | View Log