/dports/www/firefox/firefox-99.0/third_party/rust/rust_decimal/src/ops/ |
H A D | add.rs | 115 low64: lhs.low64.wrapping_sub(rhs.low64), in aligned_add() 120 if result.low64 > lhs.low64 { in aligned_add() 134 low64: lhs.low64.wrapping_add(rhs.low64), in aligned_add() 139 if result.low64 < lhs.low64 { in aligned_add() 164 result.low64 = low64; in flip_sign() 169 let mut low64 = result.low64; in reduce_scale() localVariable 190 result.low64 = low64; in reduce_scale() 206 let mut low64 = lhs.low64; in unaligned_add() localVariable 240 lhs.low64 = low64; in unaligned_add() 269 lhs.low64 = low64; in unaligned_add() [all …]
|
H A D | div.rs | 44 let low64 = self.low64(); in div32() localVariable 45 if low64 == 0 { in div32() 88 let low64 = self.low64(); in partial_divide_64() localVariable 89 if low64 < divisor { in partial_divide_64() 106 let mut low64 = self.low64(); in partial_divide_64() localVariable 107 low64 = low64.wrapping_sub(divisor << 32).wrapping_add(divisor); in partial_divide_64() 116 low64 = low64.wrapping_add(divisor); in partial_divide_64() 190 prod1 = divisor.low64(); in partial_divide_96() 625 let low64 = num.low64().wrapping_add(1); in round_up() localVariable 626 num.set_low64(low64); in round_up() [all …]
|
H A D | common.rs | 12 data: [value.low64 as u32, (value.low64 >> 32) as u32, value.hi], in from_dec64() 48 pub const fn low64(&self) -> u64 { in low64() method 86 let low64 = self.low64(); in find_scale() localVariable 142 … if hi == POWER_OVERFLOW_VALUES[x - 1].data[2] && low64 > POWER_OVERFLOW_VALUES[x - 1].low64() { in find_scale() 188 pub low64: u64, field 199 low64: m[0] as u64, in new() 213 self.low64 as u32 in lo() 217 (self.low64 >> 32) as u32 in mid() 227 self.low64 as u32, in to_decimal() 228 (self.low64 >> 32) as u32, in to_decimal() [all …]
|
H A D | rem.rs | 26 low64: d2.low64(), in rem_impl() 103 d1.low64 = quotient.low64(); in rem_impl() 110 d1.low64 %= d2.low64(); in rem_impl() 142 buffer.set_low64(d1.low64 << shift); in rem_full() 185 let divisor = d2.low64() << shift; in rem_full() 224 let low64 = buffer.low64() >> shift; in rem_full() localVariable 226 low64 as u32, in rem_full() 227 (low64 >> 32) as u32, in rem_full() 276 let low64 = (buffer.low64() >> shift) + ((buffer.data[2] as u64) << (32 - shift) << 32); in rem_full() localVariable 278 low64 as u32, in rem_full() [all …]
|
H A D | cmp.rs | 47 let mut d1_low = d1.low64; in cmp_internal() 49 let mut d2_low = d2.low64; in cmp_internal() 73 fn rescale(low64: &mut u64, high: &mut u32, diff: u32) -> bool { in rescale() 82 let tmp_lo_32 = (*low64 & U32_MASK) * power; in rescale() 83 let mut tmp = (*low64 >> 32) * power + (tmp_lo_32 >> 32); in rescale() 84 *low64 = (tmp_lo_32 & U32_MASK) + (tmp << 32); in rescale()
|
H A D | mul.rs | 20 let mut low64 = d1.lo() as u64 * d2.lo() as u64; in mul_impl() localVariable 32 let tmp = low64 / power; in mul_impl() 33 let remainder = low64 - tmp * power; in mul_impl() 34 low64 = tmp; in mul_impl() 38 if remainder >= power && (remainder > power || (low64 as u32 & 1) > 0) { in mul_impl() 39 low64 += 1; in mul_impl() 47 low64 as u32, in mul_impl() 48 (low64 >> 32) as u32, in mul_impl()
|
/dports/mail/thunderbird/thunderbird-91.8.0/third_party/rust/rust_decimal/src/ops/ |
H A D | add.rs | 115 low64: lhs.low64.wrapping_sub(rhs.low64), in aligned_add() 120 if result.low64 > lhs.low64 { in aligned_add() 134 low64: lhs.low64.wrapping_add(rhs.low64), in aligned_add() 139 if result.low64 < lhs.low64 { in aligned_add() 164 result.low64 = low64; in flip_sign() 169 let mut low64 = result.low64; in reduce_scale() localVariable 190 result.low64 = low64; in reduce_scale() 206 let mut low64 = lhs.low64; in unaligned_add() localVariable 240 lhs.low64 = low64; in unaligned_add() 269 lhs.low64 = low64; in unaligned_add() [all …]
|
H A D | div.rs | 44 let low64 = self.low64(); in div32() localVariable 45 if low64 == 0 { in div32() 88 let low64 = self.low64(); in partial_divide_64() localVariable 89 if low64 < divisor { in partial_divide_64() 106 let mut low64 = self.low64(); in partial_divide_64() localVariable 107 low64 = low64.wrapping_sub(divisor << 32).wrapping_add(divisor); in partial_divide_64() 116 low64 = low64.wrapping_add(divisor); in partial_divide_64() 190 prod1 = divisor.low64(); in partial_divide_96() 625 let low64 = num.low64().wrapping_add(1); in round_up() localVariable 626 num.set_low64(low64); in round_up() [all …]
|
H A D | common.rs | 12 data: [value.low64 as u32, (value.low64 >> 32) as u32, value.hi], in from_dec64() 48 pub const fn low64(&self) -> u64 { in low64() method 86 let low64 = self.low64(); in find_scale() localVariable 142 … if hi == POWER_OVERFLOW_VALUES[x - 1].data[2] && low64 > POWER_OVERFLOW_VALUES[x - 1].low64() { in find_scale() 188 pub low64: u64, field 199 low64: m[0] as u64, in new() 213 self.low64 as u32 in lo() 217 (self.low64 >> 32) as u32 in mid() 227 self.low64 as u32, in to_decimal() 228 (self.low64 >> 32) as u32, in to_decimal() [all …]
|
H A D | rem.rs | 26 low64: d2.low64(), in rem_impl() 103 d1.low64 = quotient.low64(); in rem_impl() 110 d1.low64 %= d2.low64(); in rem_impl() 142 buffer.set_low64(d1.low64 << shift); in rem_full() 187 let divisor = d2.low64() << shift; in rem_full() 226 let low64 = buffer.low64() >> shift; in rem_full() localVariable 228 low64 as u32, in rem_full() 229 (low64 >> 32) as u32, in rem_full() 278 let low64 = (buffer.low64() >> shift) + ((buffer.data[2] as u64) << (32 - shift) << 32); in rem_full() localVariable 280 low64 as u32, in rem_full() [all …]
|
H A D | cmp.rs | 47 let mut d1_low = d1.low64; in cmp_internal() 49 let mut d2_low = d2.low64; in cmp_internal() 73 fn rescale(low64: &mut u64, high: &mut u32, diff: u32) -> bool { in rescale() 82 let tmp_lo_32 = (*low64 & U32_MASK) * power; in rescale() 83 let mut tmp = (*low64 >> 32) * power + (tmp_lo_32 >> 32); in rescale() 84 *low64 = (tmp_lo_32 & U32_MASK) + (tmp << 32); in rescale()
|
H A D | mul.rs | 20 let mut low64 = d1.lo() as u64 * d2.lo() as u64; in mul_impl() localVariable 32 let tmp = low64 / power; in mul_impl() 33 let remainder = low64 - tmp * power; in mul_impl() 34 low64 = tmp; in mul_impl() 38 if remainder >= power && (remainder > power || (low64 as u32 & 1) > 0) { in mul_impl() 39 low64 += 1; in mul_impl() 47 low64 as u32, in mul_impl() 48 (low64 >> 32) as u32, in mul_impl()
|
/dports/www/firefox-esr/firefox-91.8.0/third_party/rust/rust_decimal/src/ops/ |
H A D | add.rs | 115 low64: lhs.low64.wrapping_sub(rhs.low64), in aligned_add() 120 if result.low64 > lhs.low64 { in aligned_add() 134 low64: lhs.low64.wrapping_add(rhs.low64), in aligned_add() 139 if result.low64 < lhs.low64 { in aligned_add() 164 result.low64 = low64; in flip_sign() 169 let mut low64 = result.low64; in reduce_scale() localVariable 190 result.low64 = low64; in reduce_scale() 206 let mut low64 = lhs.low64; in unaligned_add() localVariable 240 lhs.low64 = low64; in unaligned_add() 269 lhs.low64 = low64; in unaligned_add() [all …]
|
H A D | div.rs | 44 let low64 = self.low64(); in div32() localVariable 45 if low64 == 0 { in div32() 88 let low64 = self.low64(); in partial_divide_64() localVariable 89 if low64 < divisor { in partial_divide_64() 106 let mut low64 = self.low64(); in partial_divide_64() localVariable 107 low64 = low64.wrapping_sub(divisor << 32).wrapping_add(divisor); in partial_divide_64() 116 low64 = low64.wrapping_add(divisor); in partial_divide_64() 190 prod1 = divisor.low64(); in partial_divide_96() 625 let low64 = num.low64().wrapping_add(1); in round_up() localVariable 626 num.set_low64(low64); in round_up() [all …]
|
H A D | common.rs | 12 data: [value.low64 as u32, (value.low64 >> 32) as u32, value.hi], in from_dec64() 48 pub const fn low64(&self) -> u64 { in low64() method 86 let low64 = self.low64(); in find_scale() localVariable 142 … if hi == POWER_OVERFLOW_VALUES[x - 1].data[2] && low64 > POWER_OVERFLOW_VALUES[x - 1].low64() { in find_scale() 188 pub low64: u64, field 199 low64: m[0] as u64, in new() 213 self.low64 as u32 in lo() 217 (self.low64 >> 32) as u32 in mid() 227 self.low64 as u32, in to_decimal() 228 (self.low64 >> 32) as u32, in to_decimal() [all …]
|
H A D | rem.rs | 26 low64: d2.low64(), in rem_impl() 103 d1.low64 = quotient.low64(); in rem_impl() 110 d1.low64 %= d2.low64(); in rem_impl() 142 buffer.set_low64(d1.low64 << shift); in rem_full() 187 let divisor = d2.low64() << shift; in rem_full() 226 let low64 = buffer.low64() >> shift; in rem_full() localVariable 228 low64 as u32, in rem_full() 229 (low64 >> 32) as u32, in rem_full() 278 let low64 = (buffer.low64() >> shift) + ((buffer.data[2] as u64) << (32 - shift) << 32); in rem_full() localVariable 280 low64 as u32, in rem_full() [all …]
|
H A D | cmp.rs | 47 let mut d1_low = d1.low64; in cmp_internal() 49 let mut d2_low = d2.low64; in cmp_internal() 73 fn rescale(low64: &mut u64, high: &mut u32, diff: u32) -> bool { in rescale() 82 let tmp_lo_32 = (*low64 & U32_MASK) * power; in rescale() 83 let mut tmp = (*low64 >> 32) * power + (tmp_lo_32 >> 32); in rescale() 84 *low64 = (tmp_lo_32 & U32_MASK) + (tmp << 32); in rescale()
|
H A D | mul.rs | 20 let mut low64 = d1.lo() as u64 * d2.lo() as u64; in mul_impl() localVariable 32 let tmp = low64 / power; in mul_impl() 33 let remainder = low64 - tmp * power; in mul_impl() 34 low64 = tmp; in mul_impl() 38 if remainder >= power && (remainder > power || (low64 as u32 & 1) > 0) { in mul_impl() 39 low64 += 1; in mul_impl() 47 low64 as u32, in mul_impl() 48 (low64 >> 32) as u32, in mul_impl()
|
/dports/benchmarks/hyperfine/hyperfine-1.12.0/cargo-crates/rust_decimal-1.16.0/src/ops/ |
H A D | add.rs | 115 low64: lhs.low64.wrapping_sub(rhs.low64), 120 if result.low64 > lhs.low64 { 134 low64: lhs.low64.wrapping_add(rhs.low64), 139 if result.low64 < lhs.low64 { 164 result.low64 = low64; 169 let mut low64 = result.low64; 190 result.low64 = low64; 206 let mut low64 = lhs.low64; 240 lhs.low64 = low64; 269 lhs.low64 = low64; [all …]
|
H A D | div.rs | 44 let low64 = self.low64(); in div32() localVariable 45 if low64 == 0 { in div32() 88 let low64 = self.low64(); in partial_divide_64() localVariable 89 if low64 < divisor { in partial_divide_64() 106 let mut low64 = self.low64(); in partial_divide_64() localVariable 107 low64 = low64.wrapping_sub(divisor << 32).wrapping_add(divisor); in partial_divide_64() 116 low64 = low64.wrapping_add(divisor); in partial_divide_64() 190 prod1 = divisor.low64(); in partial_divide_96() 625 let low64 = num.low64().wrapping_add(1); in round_up() localVariable 626 num.set_low64(low64); in round_up() [all …]
|
H A D | common.rs | 12 data: [value.low64 as u32, (value.low64 >> 32) as u32, value.hi], in from_dec64() 48 pub const fn low64(&self) -> u64 { in low64() method 86 let low64 = self.low64(); in find_scale() localVariable 142 … if hi == POWER_OVERFLOW_VALUES[x - 1].data[2] && low64 > POWER_OVERFLOW_VALUES[x - 1].low64() { in find_scale() 188 pub low64: u64, field 199 low64: m[0] as u64, in new() 213 self.low64 as u32 in lo() 217 (self.low64 >> 32) as u32 in mid() 227 self.low64 as u32, in to_decimal() 228 (self.low64 >> 32) as u32, in to_decimal() [all …]
|
H A D | rem.rs | 26 low64: d2.low64(), in rem_impl() 103 d1.low64 = quotient.low64(); in rem_impl() 110 d1.low64 %= d2.low64(); in rem_impl() 142 buffer.set_low64(d1.low64 << shift); in rem_full() 185 let divisor = d2.low64() << shift; in rem_full() 224 let low64 = buffer.low64() >> shift; in rem_full() localVariable 226 low64 as u32, in rem_full() 227 (low64 >> 32) as u32, in rem_full() 276 let low64 = (buffer.low64() >> shift) + ((buffer.data[2] as u64) << (32 - shift) << 32); in rem_full() localVariable 278 low64 as u32, in rem_full() [all …]
|
H A D | cmp.rs | 47 let mut d1_low = d1.low64; in cmp_internal() 49 let mut d2_low = d2.low64; in cmp_internal() 73 fn rescale(low64: &mut u64, high: &mut u32, diff: u32) -> bool { in rescale() 82 let tmp_lo_32 = (*low64 & U32_MASK) * power; in rescale() 83 let mut tmp = (*low64 >> 32) * power + (tmp_lo_32 >> 32); in rescale() 84 *low64 = (tmp_lo_32 & U32_MASK) + (tmp << 32); in rescale()
|
H A D | mul.rs | 20 let mut low64 = d1.lo() as u64 * d2.lo() as u64; in mul_impl() localVariable 32 let tmp = low64 / power; in mul_impl() 33 let remainder = low64 - tmp * power; in mul_impl() 34 low64 = tmp; in mul_impl() 38 if remainder >= power && (remainder > power || (low64 as u32 & 1) > 0) { in mul_impl() 39 low64 += 1; in mul_impl() 47 low64 as u32, in mul_impl() 48 (low64 >> 32) as u32, in mul_impl()
|
/dports/lang/mono/mono-5.10.1.57/external/corert/src/System.Private.CoreLib/src/System/ |
H A D | Decimal.DecCalc.cs | 1095 low64 = d2.Low64; in DecAddSub() 1103 low64 = tmp64 - low64; in DecAddSub() 1189 low64 = (ulong)-(long)low64; in DecAddSub() 1190 if (low64 == 0) in DecAddSub() 1210 low64 = div; in DecAddSub() 1211 low64 <<= 32; in DecAddSub() 1213 low64 += div; in DecAddSub() 1272 d1.Low64 = low64; in DecAddSub() 1587 low64 = div; in VarDecMul() 1594 low64++; in VarDecMul() [all …]
|