/* * Licensed to the Apache Software Foundation (ASF) under one or more * contributor license agreements. See the NOTICE file distributed with * this work for additional information regarding copyright ownership. * The ASF licenses this file to You under the Apache License, Version 2.0 * (the "License"); you may not use this file except in compliance with * the License. You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ #include <string.h> #include "cbigint.h" #if defined(LINUX) || defined(FREEBSD) #define USE_LL #endif #ifdef HY_LITTLE_ENDIAN #define at(i) (i) #else #define at(i) ((i)^1) /* the sequence for halfAt is -1, 2, 1, 4, 3, 6, 5, 8... */ /* and it should correspond to 0, 1, 2, 3, 4, 5, 6, 7... */ #define halfAt(i) (-((-(i)) ^ 1)) #endif #define HIGH_IN_U64(u64) ((u64) >> 32) #if defined(USE_LL) #define LOW_IN_U64(u64) ((u64) & 0x00000000FFFFFFFFLL) #else #if defined(USE_L) #define LOW_IN_U64(u64) ((u64) & 0x00000000FFFFFFFFL) #else #define LOW_IN_U64(u64) ((u64) & 0x00000000FFFFFFFF) #endif /* USE_L */ #endif /* USE_LL */ #if defined(USE_LL) #define TEN_E1 (0xALL) #define TEN_E2 (0x64LL) #define TEN_E3 (0x3E8LL) #define TEN_E4 (0x2710LL) #define TEN_E5 (0x186A0LL) #define TEN_E6 (0xF4240LL) #define TEN_E7 (0x989680LL) #define TEN_E8 (0x5F5E100LL) #define TEN_E9 (0x3B9ACA00LL) #define TEN_E19 (0x8AC7230489E80000LL) #else #if defined(USE_L) #define TEN_E1 (0xAL) #define TEN_E2 (0x64L) #define TEN_E3 (0x3E8L) #define TEN_E4 (0x2710L) #define TEN_E5 (0x186A0L) #define TEN_E6 (0xF4240L) #define TEN_E7 (0x989680L) #define TEN_E8 (0x5F5E100L) #define TEN_E9 (0x3B9ACA00L) #define TEN_E19 (0x8AC7230489E80000L) #else #define TEN_E1 (0xA) #define TEN_E2 (0x64) #define TEN_E3 (0x3E8) #define TEN_E4 (0x2710) #define TEN_E5 (0x186A0) #define TEN_E6 (0xF4240) #define TEN_E7 (0x989680) #define TEN_E8 (0x5F5E100) #define TEN_E9 (0x3B9ACA00) #define TEN_E19 (0x8AC7230489E80000) #endif /* USE_L */ #endif /* USE_LL */ #define TIMES_TEN(x) (((x) << 3) + ((x) << 1)) #define bitSection(x, mask, shift) (((x) & (mask)) >> (shift)) #define DOUBLE_TO_LONGBITS(dbl) (*((U_64 *)(&dbl))) #define FLOAT_TO_INTBITS(flt) (*((U_32 *)(&flt))) #define CREATE_DOUBLE_BITS(normalizedM, e) (((normalizedM) & MANTISSA_MASK) | (((U_64)((e) + E_OFFSET)) << 52)) #if defined(USE_LL) #define MANTISSA_MASK (0x000FFFFFFFFFFFFFLL) #define EXPONENT_MASK (0x7FF0000000000000LL) #define NORMAL_MASK (0x0010000000000000LL) #define SIGN_MASK (0x8000000000000000LL) #else #if defined(USE_L) #define MANTISSA_MASK (0x000FFFFFFFFFFFFFL) #define EXPONENT_MASK (0x7FF0000000000000L) #define NORMAL_MASK (0x0010000000000000L) #define SIGN_MASK (0x8000000000000000L) #else #define MANTISSA_MASK (0x000FFFFFFFFFFFFF) #define EXPONENT_MASK (0x7FF0000000000000) #define NORMAL_MASK (0x0010000000000000) #define SIGN_MASK (0x8000000000000000) #endif /* USE_L */ #endif /* USE_LL */ #define E_OFFSET (1075) #define FLOAT_MANTISSA_MASK (0x007FFFFF) #define FLOAT_EXPONENT_MASK (0x7F800000) #define FLOAT_NORMAL_MASK (0x00800000) #define FLOAT_E_OFFSET (150) IDATA simpleAddHighPrecision (U_64 * arg1, IDATA length, U_64 arg2) { /* assumes length > 0 */ IDATA index = 1; *arg1 += arg2; if (arg2 <= *arg1) return 0; else if (length == 1) return 1; while (++arg1[index] == 0 && ++index < length); return (IDATA) index == length; } IDATA addHighPrecision (U_64 * arg1, IDATA length1, U_64 * arg2, IDATA length2) { /* addition is limited by length of arg1 as it this function is * storing the result in arg1 */ /* fix for cc (GCC) 3.2 20020903 (Red Hat Linux 8.0 3.2-7): code generated does not * do the temp1 + temp2 + carry addition correct. carry is 64 bit because gcc has * subtle issues when you mix 64 / 32 bit maths. */ U_64 temp1, temp2, temp3; /* temporary variables to help the SH-4, and gcc */ U_64 carry; IDATA index; if (length1 == 0 || length2 == 0) { return 0; } else if (length1 < length2) { length2 = length1; } carry = 0; index = 0; do { temp1 = arg1[index]; temp2 = arg2[index]; temp3 = temp1 + temp2; arg1[index] = temp3 + carry; if (arg2[index] < arg1[index]) carry = 0; else if (arg2[index] != arg1[index]) carry = 1; } while (++index < length2); if (!carry) return 0; else if (index == length1) return 1; while (++arg1[index] == 0 && ++index < length1); return (IDATA) index == length1; } void subtractHighPrecision (U_64 * arg1, IDATA length1, U_64 * arg2, IDATA length2) { /* assumes arg1 > arg2 */ IDATA index; for (index = 0; index < length1; ++index) arg1[index] = ~arg1[index]; simpleAddHighPrecision (arg1, length1, 1); while (length2 > 0 && arg2[length2 - 1] == 0) --length2; addHighPrecision (arg1, length1, arg2, length2); for (index = 0; index < length1; ++index) arg1[index] = ~arg1[index]; simpleAddHighPrecision (arg1, length1, 1); } U_32 simpleMultiplyHighPrecision (U_64 * arg1, IDATA length, U_64 arg2) { /* assumes arg2 only holds 32 bits of information */ U_64 product; IDATA index; index = 0; product = 0; do { product = HIGH_IN_U64 (product) + arg2 * LOW_U32_FROM_PTR (arg1 + index); LOW_U32_FROM_PTR (arg1 + index) = LOW_U32_FROM_VAR (product); product = HIGH_IN_U64 (product) + arg2 * HIGH_U32_FROM_PTR (arg1 + index); HIGH_U32_FROM_PTR (arg1 + index) = LOW_U32_FROM_VAR (product); } while (++index < length); return HIGH_U32_FROM_VAR (product); } void simpleMultiplyAddHighPrecision (U_64 * arg1, IDATA length, U_64 arg2, U_32 * result) { /* Assumes result can hold the product and arg2 only holds 32 bits of information */ U_64 product; IDATA index, resultIndex; index = resultIndex = 0; product = 0; do { product = HIGH_IN_U64 (product) + result[at (resultIndex)] + arg2 * LOW_U32_FROM_PTR (arg1 + index); result[at (resultIndex)] = LOW_U32_FROM_VAR (product); ++resultIndex; product = HIGH_IN_U64 (product) + result[at (resultIndex)] + arg2 * HIGH_U32_FROM_PTR (arg1 + index); result[at (resultIndex)] = LOW_U32_FROM_VAR (product); ++resultIndex; } while (++index < length); result[at (resultIndex)] += HIGH_U32_FROM_VAR (product); if (result[at (resultIndex)] < HIGH_U32_FROM_VAR (product)) { /* must be careful with ++ operator and macro expansion */ ++resultIndex; while (++result[at (resultIndex)] == 0) ++resultIndex; } } void multiplyHighPrecision (U_64 * arg1, IDATA length1, U_64 * arg2, IDATA length2, U_64 * result, IDATA length) { /* assumes result is large enough to hold product */ U_64 *temp; U_32 *resultIn32; IDATA count, index; if (length1 < length2) { temp = arg1; arg1 = arg2; arg2 = temp; count = length1; length1 = length2; length2 = count; } memset (result, 0, sizeof (U_64) * length); /* length1 > length2 */ resultIn32 = (U_32 *) result; index = -1; for (count = 0; count < length2; ++count) { simpleMultiplyAddHighPrecision (arg1, length1, LOW_IN_U64 (arg2[count]), resultIn32 + (++index)); simpleMultiplyAddHighPrecision (arg1, length1, HIGH_IN_U64 (arg2[count]), resultIn32 + (++index)); } } U_32 simpleAppendDecimalDigitHighPrecision (U_64 * arg1, IDATA length, U_64 digit) { /* assumes digit is less than 32 bits */ U_64 arg; IDATA index = 0; digit <<= 32; do { arg = LOW_IN_U64 (arg1[index]); digit = HIGH_IN_U64 (digit) + TIMES_TEN (arg); LOW_U32_FROM_PTR (arg1 + index) = LOW_U32_FROM_VAR (digit); arg = HIGH_IN_U64 (arg1[index]); digit = HIGH_IN_U64 (digit) + TIMES_TEN (arg); HIGH_U32_FROM_PTR (arg1 + index) = LOW_U32_FROM_VAR (digit); } while (++index < length); return HIGH_U32_FROM_VAR (digit); } void simpleShiftLeftHighPrecision (U_64 * arg1, IDATA length, IDATA arg2) { /* assumes length > 0 */ IDATA index, offset; if (arg2 >= 64) { offset = arg2 >> 6; index = length; while (--index - offset >= 0) arg1[index] = arg1[index - offset]; do { arg1[index] = 0; } while (--index >= 0); arg2 &= 0x3F; } if (arg2 == 0) return; while (--length > 0) { arg1[length] = arg1[length] << arg2 | arg1[length - 1] >> (64 - arg2); } *arg1 <<= arg2; } IDATA highestSetBit (U_64 * y) { U_32 x; IDATA result; if (*y == 0) return 0; #if defined(USE_LL) if (*y & 0xFFFFFFFF00000000LL) { x = HIGH_U32_FROM_PTR (y); result = 32; } else { x = LOW_U32_FROM_PTR (y); result = 0; } #else #if defined(USE_L) if (*y & 0xFFFFFFFF00000000L) { x = HIGH_U32_FROM_PTR (y); result = 32; } else { x = LOW_U32_FROM_PTR (y); result = 0; } #else if (*y & 0xFFFFFFFF00000000) { x = HIGH_U32_FROM_PTR (y); result = 32; } else { x = LOW_U32_FROM_PTR (y); result = 0; } #endif /* USE_L */ #endif /* USE_LL */ if (x & 0xFFFF0000) { x = bitSection (x, 0xFFFF0000, 16); result += 16; } if (x & 0xFF00) { x = bitSection (x, 0xFF00, 8); result += 8; } if (x & 0xF0) { x = bitSection (x, 0xF0, 4); result += 4; } if (x > 0x7) return result + 4; else if (x > 0x3) return result + 3; else if (x > 0x1) return result + 2; else return result + 1; } IDATA lowestSetBit (U_64 * y) { U_32 x; IDATA result; if (*y == 0) return 0; #if defined(USE_LL) if (*y & 0x00000000FFFFFFFFLL) { x = LOW_U32_FROM_PTR (y); result = 0; } else { x = HIGH_U32_FROM_PTR (y); result = 32; } #else #if defined(USE_L) if (*y & 0x00000000FFFFFFFFL) { x = LOW_U32_FROM_PTR (y); result = 0; } else { x = HIGH_U32_FROM_PTR (y); result = 32; } #else if (*y & 0x00000000FFFFFFFF) { x = LOW_U32_FROM_PTR (y); result = 0; } else { x = HIGH_U32_FROM_PTR (y); result = 32; } #endif /* USE_L */ #endif /* USE_LL */ if (!(x & 0xFFFF)) { x = bitSection (x, 0xFFFF0000, 16); result += 16; } if (!(x & 0xFF)) { x = bitSection (x, 0xFF00, 8); result += 8; } if (!(x & 0xF)) { x = bitSection (x, 0xF0, 4); result += 4; } if (x & 0x1) return result + 1; else if (x & 0x2) return result + 2; else if (x & 0x4) return result + 3; else return result + 4; } IDATA highestSetBitHighPrecision (U_64 * arg, IDATA length) { IDATA highBit; while (--length >= 0) { highBit = highestSetBit (arg + length); if (highBit) return highBit + 64 * length; } return 0; } IDATA lowestSetBitHighPrecision (U_64 * arg, IDATA length) { IDATA lowBit, index = -1; while (++index < length) { lowBit = lowestSetBit (arg + index); if (lowBit) return lowBit + 64 * index; } return 0; } IDATA compareHighPrecision (U_64 * arg1, IDATA length1, U_64 * arg2, IDATA length2) { while (--length1 >= 0 && arg1[length1] == 0); while (--length2 >= 0 && arg2[length2] == 0); if (length1 > length2) return 1; else if (length1 < length2) return -1; else if (length1 > -1) { do { if (arg1[length1] > arg2[length1]) return 1; else if (arg1[length1] < arg2[length1]) return -1; } while (--length1 >= 0); } return 0; } jdouble toDoubleHighPrecision (U_64 * arg, IDATA length) { IDATA highBit; U_64 mantissa, test64; U_32 test; jdouble result; while (length > 0 && arg[length - 1] == 0) --length; if (length == 0) result = 0.0; else if (length > 16) { DOUBLE_TO_LONGBITS (result) = EXPONENT_MASK; } else if (length == 1) { highBit = highestSetBit (arg); if (highBit <= 53) { highBit = 53 - highBit; mantissa = *arg << highBit; DOUBLE_TO_LONGBITS (result) = CREATE_DOUBLE_BITS (mantissa, -highBit); } else { highBit -= 53; mantissa = *arg >> highBit; DOUBLE_TO_LONGBITS (result) = CREATE_DOUBLE_BITS (mantissa, highBit); /* perform rounding, round to even in case of tie */ test = (LOW_U32_FROM_PTR (arg) << (11 - highBit)) & 0x7FF; if (test > 0x400 || ((test == 0x400) && (mantissa & 1))) DOUBLE_TO_LONGBITS (result) = DOUBLE_TO_LONGBITS (result) + 1; } } else { highBit = highestSetBit (arg + (--length)); if (highBit <= 53) { highBit = 53 - highBit; if (highBit > 0) { mantissa = (arg[length] << highBit) | (arg[length - 1] >> (64 - highBit)); } else { mantissa = arg[length]; } DOUBLE_TO_LONGBITS (result) = CREATE_DOUBLE_BITS (mantissa, length * 64 - highBit); /* perform rounding, round to even in case of tie */ test64 = arg[--length] << highBit; if (test64 > SIGN_MASK || ((test64 == SIGN_MASK) && (mantissa & 1))) DOUBLE_TO_LONGBITS (result) = DOUBLE_TO_LONGBITS (result) + 1; else if (test64 == SIGN_MASK) { while (--length >= 0) { if (arg[length] != 0) { DOUBLE_TO_LONGBITS (result) = DOUBLE_TO_LONGBITS (result) + 1; break; } } } } else { highBit -= 53; mantissa = arg[length] >> highBit; DOUBLE_TO_LONGBITS (result) = CREATE_DOUBLE_BITS (mantissa, length * 64 + highBit); /* perform rounding, round to even in case of tie */ test = (LOW_U32_FROM_PTR (arg + length) << (11 - highBit)) & 0x7FF; if (test > 0x400 || ((test == 0x400) && (mantissa & 1))) DOUBLE_TO_LONGBITS (result) = DOUBLE_TO_LONGBITS (result) + 1; else if (test == 0x400) { do { if (arg[--length] != 0) { DOUBLE_TO_LONGBITS (result) = DOUBLE_TO_LONGBITS (result) + 1; break; } } while (length > 0); } } } return result; } IDATA tenToTheEHighPrecision (U_64 * result, IDATA length, jint e) { /* size test */ if (length < ((e / 19) + 1)) return 0; memset (result, 0, length * sizeof (U_64)); *result = 1; if (e == 0) return 1; length = 1; length = timesTenToTheEHighPrecision (result, length, e); /* bad O(n) way of doing it, but simple */ /* do { overflow = simpleAppendDecimalDigitHighPrecision(result, length, 0); if (overflow) result[length++] = overflow; } while (--e); */ return length; } IDATA timesTenToTheEHighPrecision (U_64 * result, IDATA length, jint e) { /* assumes result can hold value */ U_64 overflow; int exp10 = e; if (e == 0) return length; /* bad O(n) way of doing it, but simple */ /* do { overflow = simpleAppendDecimalDigitHighPrecision(result, length, 0); if (overflow) result[length++] = overflow; } while (--e); */ /* Replace the current implementaion which performs a * "multiplication" by 10 e number of times with an actual * mulitplication. 10e19 is the largest exponent to the power of ten * that will fit in a 64-bit integer, and 10e9 is the largest exponent to * the power of ten that will fit in a 64-bit integer. Not sure where the * break-even point is between an actual multiplication and a * simpleAappendDecimalDigit() so just pick 10e3 as that point for * now. */ while (exp10 >= 19) { overflow = simpleMultiplyHighPrecision64 (result, length, TEN_E19); if (overflow) result[length++] = overflow; exp10 -= 19; } while (exp10 >= 9) { overflow = simpleMultiplyHighPrecision (result, length, TEN_E9); if (overflow) result[length++] = overflow; exp10 -= 9; } if (exp10 == 0) return length; else if (exp10 == 1) { overflow = simpleAppendDecimalDigitHighPrecision (result, length, 0); if (overflow) result[length++] = overflow; } else if (exp10 == 2) { overflow = simpleAppendDecimalDigitHighPrecision (result, length, 0); if (overflow) result[length++] = overflow; overflow = simpleAppendDecimalDigitHighPrecision (result, length, 0); if (overflow) result[length++] = overflow; } else if (exp10 == 3) { overflow = simpleMultiplyHighPrecision (result, length, TEN_E3); if (overflow) result[length++] = overflow; } else if (exp10 == 4) { overflow = simpleMultiplyHighPrecision (result, length, TEN_E4); if (overflow) result[length++] = overflow; } else if (exp10 == 5) { overflow = simpleMultiplyHighPrecision (result, length, TEN_E5); if (overflow) result[length++] = overflow; } else if (exp10 == 6) { overflow = simpleMultiplyHighPrecision (result, length, TEN_E6); if (overflow) result[length++] = overflow; } else if (exp10 == 7) { overflow = simpleMultiplyHighPrecision (result, length, TEN_E7); if (overflow) result[length++] = overflow; } else if (exp10 == 8) { overflow = simpleMultiplyHighPrecision (result, length, TEN_E8); if (overflow) result[length++] = overflow; } return length; } U_64 doubleMantissa (jdouble z) { U_64 m = DOUBLE_TO_LONGBITS (z); if ((m & EXPONENT_MASK) != 0) m = (m & MANTISSA_MASK) | NORMAL_MASK; else m = (m & MANTISSA_MASK); return m; } IDATA doubleExponent (jdouble z) { /* assumes positive double */ IDATA k = HIGH_U32_FROM_VAR (z) >> 20; if (k) k -= E_OFFSET; else k = 1 - E_OFFSET; return k; } UDATA floatMantissa (jfloat z) { UDATA m = (UDATA) FLOAT_TO_INTBITS (z); if ((m & FLOAT_EXPONENT_MASK) != 0) m = (m & FLOAT_MANTISSA_MASK) | FLOAT_NORMAL_MASK; else m = (m & FLOAT_MANTISSA_MASK); return m; } IDATA floatExponent (jfloat z) { /* assumes positive float */ IDATA k = FLOAT_TO_INTBITS (z) >> 23; if (k) k -= FLOAT_E_OFFSET; else k = 1 - FLOAT_E_OFFSET; return k; } /* Allow a 64-bit value in arg2 */ U_64 simpleMultiplyHighPrecision64 (U_64 * arg1, IDATA length, U_64 arg2) { U_64 intermediate, *pArg1, carry1, carry2, prod1, prod2, sum; IDATA index; U_32 buf32; index = 0; intermediate = 0; pArg1 = arg1 + index; carry1 = carry2 = 0; do { if ((*pArg1 != 0) || (intermediate != 0)) { prod1 = (U_64) LOW_U32_FROM_VAR (arg2) * (U_64) LOW_U32_FROM_PTR (pArg1); sum = intermediate + prod1; if ((sum < prod1) || (sum < intermediate)) { carry1 = 1; } else { carry1 = 0; } prod1 = (U_64) LOW_U32_FROM_VAR (arg2) * (U_64) HIGH_U32_FROM_PTR (pArg1); prod2 = (U_64) HIGH_U32_FROM_VAR (arg2) * (U_64) LOW_U32_FROM_PTR (pArg1); intermediate = carry2 + HIGH_IN_U64 (sum) + prod1 + prod2; if ((intermediate < prod1) || (intermediate < prod2)) { carry2 = 1; } else { carry2 = 0; } LOW_U32_FROM_PTR (pArg1) = LOW_U32_FROM_VAR (sum); buf32 = HIGH_U32_FROM_PTR (pArg1); HIGH_U32_FROM_PTR (pArg1) = LOW_U32_FROM_VAR (intermediate); intermediate = carry1 + HIGH_IN_U64 (intermediate) + (U_64) HIGH_U32_FROM_VAR (arg2) * (U_64) buf32; } pArg1++; } while (++index < length); return intermediate; }