Lines Matching refs:aSig0

344      bits32 aSig0,  in normalizeFloat64Subnormal()  argument
353 if ( aSig0 == 0 ) { in normalizeFloat64Subnormal()
366 shiftCount = countLeadingZeros32( aSig0 ) - 11; in normalizeFloat64Subnormal()
367 shortShift64Left( aSig0, aSig1, shiftCount, zSig0Ptr, zSig1Ptr ); in normalizeFloat64Subnormal()
1419 bits32 aSig0, aSig1, absZ, aSigExtra; in float64_to_int32() local
1424 aSig0 = extractFloat64Frac0( a ); in float64_to_int32()
1430 if ( ( aExp == 0x7FF ) && ( aSig0 | aSig1 ) ) aSign = 0; in float64_to_int32()
1434 aSig0 | 0x00100000, aSig1, shiftCount, &absZ, &aSigExtra ); in float64_to_int32()
1440 aSigExtra = aExp | aSig0 | aSig1; in float64_to_int32()
1444 aSig0 |= 0x00100000; in float64_to_int32()
1445 aSigExtra = ( aSig0<<( shiftCount & 31 ) ) | aSig1; in float64_to_int32()
1446 absZ = aSig0>>( - shiftCount ); in float64_to_int32()
1493 bits32 aSig0, aSig1, absZ, aSigExtra; in float64_to_int32_round_to_zero() local
1497 aSig0 = extractFloat64Frac0( a ); in float64_to_int32_round_to_zero()
1503 if ( ( aExp == 0x7FF ) && ( aSig0 | aSig1 ) ) aSign = 0; in float64_to_int32_round_to_zero()
1507 aSig0 | 0x00100000, aSig1, shiftCount, &absZ, &aSigExtra ); in float64_to_int32_round_to_zero()
1511 if ( aExp | aSig0 | aSig1 ) { in float64_to_int32_round_to_zero()
1516 aSig0 |= 0x00100000; in float64_to_int32_round_to_zero()
1517 aSigExtra = ( aSig0<<( shiftCount & 31 ) ) | aSig1; in float64_to_int32_round_to_zero()
1518 absZ = aSig0>>( - shiftCount ); in float64_to_int32_round_to_zero()
1543 bits32 aSig0, aSig1, zSig; in float64_to_float32() local
1547 aSig0 = extractFloat64Frac0( a ); in float64_to_float32()
1551 if ( aSig0 | aSig1 ) { in float64_to_float32()
1556 shift64RightJamming( aSig0, aSig1, 22, &allZero, &zSig ); in float64_to_float32()
1678 bits32 aSig0, aSig1, bSig0, bSig1, zSig0, zSig1, zSig2; in addFloat64Sigs() local
1682 aSig0 = extractFloat64Frac0( a ); in addFloat64Sigs()
1690 if ( aSig0 | aSig1 ) return propagateFloat64NaN( a, b ); in addFloat64Sigs()
1712 aSig0 |= 0x00100000; in addFloat64Sigs()
1715 aSig0, aSig1, 0, - expDiff, &aSig0, &aSig1, &zSig2 ); in addFloat64Sigs()
1720 if ( aSig0 | aSig1 | bSig0 | bSig1 ) { in addFloat64Sigs()
1725 add64( aSig0, aSig1, bSig0, bSig1, &zSig0, &zSig1 ); in addFloat64Sigs()
1732 aSig0 |= 0x00100000; in addFloat64Sigs()
1733 add64( aSig0, aSig1, bSig0, bSig1, &zSig0, &zSig1 ); in addFloat64Sigs()
1756 bits32 aSig0, aSig1, bSig0, bSig1, zSig0, zSig1; in subFloat64Sigs() local
1760 aSig0 = extractFloat64Frac0( a ); in subFloat64Sigs()
1766 shortShift64Left( aSig0, aSig1, 10, &aSig0, &aSig1 ); in subFloat64Sigs()
1771 if ( aSig0 | aSig1 | bSig0 | bSig1 ) { in subFloat64Sigs()
1781 if ( bSig0 < aSig0 ) goto aBigger; in subFloat64Sigs()
1782 if ( aSig0 < bSig0 ) goto bBigger; in subFloat64Sigs()
1795 aSig0 |= 0x40000000; in subFloat64Sigs()
1797 shift64RightJamming( aSig0, aSig1, - expDiff, &aSig0, &aSig1 ); in subFloat64Sigs()
1800 sub64( bSig0, bSig1, aSig0, aSig1, &zSig0, &zSig1 ); in subFloat64Sigs()
1806 if ( aSig0 | aSig1 ) return propagateFloat64NaN( a, b ); in subFloat64Sigs()
1816 aSig0 |= 0x40000000; in subFloat64Sigs()
1818 sub64( aSig0, aSig1, bSig0, bSig1, &zSig0, &zSig1 ); in subFloat64Sigs()
1881 bits32 aSig0, aSig1, bSig0, bSig1, zSig0, zSig1, zSig2, zSig3; in float64_mul() local
1884 aSig0 = extractFloat64Frac0( a ); in float64_mul()
1893 if ( ( aSig0 | aSig1 ) in float64_mul()
1902 if ( ( aExp | aSig0 | aSig1 ) == 0 ) { in float64_mul()
1910 if ( ( aSig0 | aSig1 ) == 0 ) return packFloat64( zSign, 0, 0, 0 ); in float64_mul()
1911 normalizeFloat64Subnormal( aSig0, aSig1, &aExp, &aSig0, &aSig1 ); in float64_mul()
1918 aSig0 |= 0x00100000; in float64_mul()
1920 mul64To128( aSig0, aSig1, bSig0, bSig1, &zSig0, &zSig1, &zSig2, &zSig3 ); in float64_mul()
1921 add64( zSig0, zSig1, aSig0, aSig1, &zSig0, &zSig1 ); in float64_mul()
1943 bits32 aSig0, aSig1, bSig0, bSig1, zSig0, zSig1, zSig2; in float64_div() local
1947 aSig0 = extractFloat64Frac0( a ); in float64_div()
1956 if ( aSig0 | aSig1 ) return propagateFloat64NaN( a, b ); in float64_div()
1969 if ( ( aExp | aSig0 | aSig1 ) == 0 ) { in float64_div()
1980 if ( ( aSig0 | aSig1 ) == 0 ) return packFloat64( zSign, 0, 0, 0 ); in float64_div()
1981 normalizeFloat64Subnormal( aSig0, aSig1, &aExp, &aSig0, &aSig1 ); in float64_div()
1984 shortShift64Left( aSig0 | 0x00100000, aSig1, 11, &aSig0, &aSig1 ); in float64_div()
1986 if ( le64( bSig0, bSig1, aSig0, aSig1 ) ) { in float64_div()
1987 shift64Right( aSig0, aSig1, 1, &aSig0, &aSig1 ); in float64_div()
1990 zSig0 = estimateDiv64To32( aSig0, aSig1, bSig0 ); in float64_div()
1992 sub96( aSig0, aSig1, 0, term0, term1, term2, &rem0, &rem1, &rem2 ); in float64_div()
2024 bits32 aSig0, aSig1, bSig0, bSig1, q, term0, term1, term2; in float64_rem() local
2030 aSig0 = extractFloat64Frac0( a ); in float64_rem()
2038 if ( ( aSig0 | aSig1 ) in float64_rem()
2057 if ( ( aSig0 | aSig1 ) == 0 ) return a; in float64_rem()
2058 normalizeFloat64Subnormal( aSig0, aSig1, &aExp, &aSig0, &aSig1 ); in float64_rem()
2063 aSig0 | 0x00100000, aSig1, 11 - ( expDiff < 0 ), &aSig0, &aSig1 ); in float64_rem()
2065 q = le64( bSig0, bSig1, aSig0, aSig1 ); in float64_rem()
2066 if ( q ) sub64( aSig0, aSig1, bSig0, bSig1, &aSig0, &aSig1 ); in float64_rem()
2069 q = estimateDiv64To32( aSig0, aSig1, bSig0 ); in float64_rem()
2073 shortShift64Left( aSig0, aSig1, 29, &aSig0, &allZero ); in float64_rem()
2074 sub64( aSig0, 0, term1, term2, &aSig0, &aSig1 ); in float64_rem()
2078 q = estimateDiv64To32( aSig0, aSig1, bSig0 ); in float64_rem()
2084 shift64Right( aSig0, aSig1, - expDiff, &aSig0, &aSig1 ); in float64_rem()
2087 shortShift64Left( aSig0, aSig1, expDiff, &aSig0, &aSig1 ); in float64_rem()
2090 sub64( aSig0, aSig1, term1, term2, &aSig0, &aSig1 ); in float64_rem()
2093 shift64Right( aSig0, aSig1, 8, &aSig0, &aSig1 ); in float64_rem()
2097 alternateASig0 = aSig0; in float64_rem()
2100 sub64( aSig0, aSig1, bSig0, bSig1, &aSig0, &aSig1 ); in float64_rem()
2101 } while ( 0 <= (sbits32) aSig0 ); in float64_rem()
2103 aSig0, aSig1, alternateASig0, alternateASig1, &sigMean0, &sigMean1 ); in float64_rem()
2106 aSig0 = alternateASig0; in float64_rem()
2109 zSign = ( (sbits32) aSig0 < 0 ); in float64_rem()
2110 if ( zSign ) sub64( 0, 0, aSig0, aSig1, &aSig0, &aSig1 ); in float64_rem()
2112 normalizeRoundAndPackFloat64( aSign ^ zSign, bExp - 4, aSig0, aSig1 ); in float64_rem()
2129 bits32 aSig0, aSig1, zSig0, zSig1, zSig2, doubleZSig0; in float64_sqrt() local
2134 aSig0 = extractFloat64Frac0( a ); in float64_sqrt()
2138 if ( aSig0 | aSig1 ) return propagateFloat64NaN( a, a ); in float64_sqrt()
2143 if ( ( aExp | aSig0 | aSig1 ) == 0 ) return a; in float64_sqrt()
2149 if ( ( aSig0 | aSig1 ) == 0 ) return packFloat64( 0, 0, 0, 0 ); in float64_sqrt()
2150 normalizeFloat64Subnormal( aSig0, aSig1, &aExp, &aSig0, &aSig1 ); in float64_sqrt()
2153 aSig0 |= 0x00100000; in float64_sqrt()
2154 shortShift64Left( aSig0, aSig1, 11, &term0, &term1 ); in float64_sqrt()
2158 shortShift64Left( aSig0, aSig1, 9 - ( aExp & 1 ), &aSig0, &aSig1 ); in float64_sqrt()
2160 sub64( aSig0, aSig1, term0, term1, &rem0, &rem1 ); in float64_sqrt()