use crate::common::dd_fmla;
use crate::double_double::DoubleDouble;
use crate::logs::log_dd::log_poly;
use crate::logs::log_dd_coeffs::LOG_NEG_DD;
use crate::polyeval::{f_estrin_polyeval7, f_polyeval3};
use crate::pow_tables::POW_INVERSE;
#[inline(always)]
pub(crate) fn log1p_tiny(z: f64) -> DoubleDouble {
const Q_1: [(u64, u64); 7] = [
(0xbc85555555555555, 0x3fd5555555555556),
(0x0000000000000000, 0xbfd0000000000000),
(0xbc6999999999999a, 0x3fc999999999999a),
(0x3c75555555555555, 0xbfc5555555555556),
(0x3c62492492492492, 0x3fc2492492492492),
(0x0000000000000000, 0xbfc0000000000000),
(0x3c5c71c71c71c71c, 0x3fbc71c71c71c71c),
];
let mut r = DoubleDouble::quick_mul_f64_add_f64(
DoubleDouble::from_bit_pair(Q_1[6]),
z,
f64::from_bits(0xbfc0000000000000),
);
r = DoubleDouble::quick_mul_f64_add(r, z, DoubleDouble::from_bit_pair(Q_1[4]));
r = DoubleDouble::quick_mul_f64_add(r, z, DoubleDouble::from_bit_pair(Q_1[3]));
r = DoubleDouble::quick_mul_f64_add(r, z, DoubleDouble::from_bit_pair(Q_1[2]));
r = DoubleDouble::quick_mul_f64_add_f64(r, z, f64::from_bits(0xbfd0000000000000));
r = DoubleDouble::quick_mul_f64_add(r, z, DoubleDouble::from_bit_pair(Q_1[0]));
r = DoubleDouble::quick_mul_f64_add_f64(r, z, f64::from_bits(0xbfe0000000000000));
r = DoubleDouble::quick_mul_f64_add_f64(r, z, f64::from_bits(0x3ff0000000000000));
DoubleDouble::quick_mult_f64(r, z)
}
#[inline(always)]
fn log1p_poly_fast(z: f64) -> DoubleDouble {
const Q: [f64; 7] = [
f64::from_bits(0x3fd5555555555556),
f64::from_bits(0xbfcfffffffffffdc),
f64::from_bits(0x3fc99999998fd488),
f64::from_bits(0xbfc5555555d90fc7),
f64::from_bits(0x3fc24936d06d3bf8),
f64::from_bits(0xbfbfff46726d8e88),
f64::from_bits(0x3fa1847e2faea348),
];
let x2 = DoubleDouble::from_exact_mult(z, z);
let p = f_estrin_polyeval7(z, Q[0], Q[1], Q[2], Q[3], Q[4], Q[5], Q[6]);
let mut t = DoubleDouble::quick_mult_f64(x2, p);
t = DoubleDouble::quick_mult_f64(t, z);
DoubleDouble::mul_f64_add(x2, -0.5, t)
}
#[inline(always)]
fn log1p_tiny_fast(z: f64) -> DoubleDouble {
let x2 = DoubleDouble::from_exact_mult(z, z);
let p = f_polyeval3(
z,
f64::from_bits(0xbfcffffffffffff4),
f64::from_bits(0x3fc99999b4d2481f),
f64::from_bits(0xbfc55555714a3cb8),
);
let DoubleDouble { hi: h, lo: r } = DoubleDouble::from_exact_mult(z, p);
let DoubleDouble { hi: ph, lo: q } = DoubleDouble::add_f64(
DoubleDouble::from_bit_pair((0xbc77e8068b994170, 0x3fd5555555555551)),
h,
);
let p = DoubleDouble::new(r + q, ph);
let mut t = DoubleDouble::quick_mult(x2, p);
t = DoubleDouble::quick_mult_f64(t, z);
let f = DoubleDouble::f64_add(z, DoubleDouble::mul_f64_add(x2, -0.5, t));
DoubleDouble::from_exact_add(f.hi, f.lo)
}
#[inline]
pub(crate) fn log1p_dd(z: f64) -> DoubleDouble {
let ax = z.to_bits().wrapping_shl(1);
if ax < 0x7e60000000000000u64 {
return log1p_tiny(z);
}
let dz = DoubleDouble::from_full_exact_add(z, 1.0);
let log_lo = if dz.hi <= f64::from_bits(0x7fd0000000000000) || dz.lo.abs() >= 4.0 {
dz.lo / dz.hi
} else {
0.
};
let x_u = dz.hi.to_bits();
let mut m = x_u & 0xfffffffffffff;
let mut e: i64 = ((x_u >> 52) & 0x7ff) as i64;
let t;
if e != 0 {
t = m | (0x3ffu64 << 52);
m = m.wrapping_add(1u64 << 52);
e -= 0x3ff;
} else {
let k = m.leading_zeros() - 11;
e = -0x3fei64 - k as i64;
m = m.wrapping_shl(k);
t = m | (0x3ffu64 << 52);
}
let mut t = f64::from_bits(t);
let c: usize = (m >= 0x16a09e667f3bcd) as usize;
static CY: [f64; 2] = [1.0, 0.5];
static CM: [u64; 2] = [44, 45];
e = e.wrapping_add(c as i64);
let be = e;
let i = m >> CM[c];
t *= CY[c];
let r = f64::from_bits(POW_INVERSE[(i - 181) as usize]);
let log_r = DoubleDouble::from_bit_pair(LOG_NEG_DD[(i - 181) as usize]);
let z = f64::mul_add(r, t, -1.0);
const LOG2_DD: DoubleDouble = DoubleDouble::new(
f64::from_bits(0x3c7abc9e3b39803f),
f64::from_bits(0x3fe62e42fefa39ef),
);
let tt = DoubleDouble::mul_f64_add(LOG2_DD, be as f64, log_r);
let v = DoubleDouble::full_add_f64(tt, z);
let mut p = log_poly(z);
p.lo += log_lo;
DoubleDouble::f64_add(v.hi, DoubleDouble::new(v.lo + p.lo, p.hi))
}
#[inline]
pub(crate) fn log1p_fast_dd(z: f64) -> DoubleDouble {
let ax = z.to_bits().wrapping_shl(1);
if ax < 0x7e60000000000000u64 {
return log1p_tiny_fast(z);
}
let dz = DoubleDouble::from_full_exact_add(z, 1.0);
let log_lo = if dz.hi <= f64::from_bits(0x7fd0000000000000) || dz.lo.abs() >= 4.0 {
dz.lo / dz.hi
} else {
0.
};
let x_u = dz.hi.to_bits();
let mut m = x_u & 0xfffffffffffff;
let mut e: i64 = ((x_u >> 52) & 0x7ff) as i64;
let t;
if e != 0 {
t = m | (0x3ffu64 << 52);
m = m.wrapping_add(1u64 << 52);
e -= 0x3ff;
} else {
let k = m.leading_zeros() - 11;
e = -0x3fei64 - k as i64;
m = m.wrapping_shl(k);
t = m | (0x3ffu64 << 52);
}
let mut t = f64::from_bits(t);
let c: usize = (m >= 0x16a09e667f3bcd) as usize;
static CY: [f64; 2] = [1.0, 0.5];
static CM: [u64; 2] = [44, 45];
e = e.wrapping_add(c as i64);
let be = e;
let i = m >> CM[c];
t *= CY[c];
let r = f64::from_bits(POW_INVERSE[(i - 181) as usize]);
let log_r = DoubleDouble::from_bit_pair(LOG_NEG_DD[(i - 181) as usize]);
let z = dd_fmla(r, t, -1.0);
const LOG2_DD: DoubleDouble = DoubleDouble::new(
f64::from_bits(0x3c7abc9e3b39803f),
f64::from_bits(0x3fe62e42fefa39ef),
);
let tt = DoubleDouble::mul_f64_add(LOG2_DD, be as f64, log_r);
let v = DoubleDouble::full_add_f64(tt, z);
let mut p = log1p_poly_fast(z);
p.lo += log_lo;
DoubleDouble::f64_add(v.hi, DoubleDouble::new(v.lo + p.lo, p.hi))
}