1
 2
 3
 4
 5
 6
 7
 8
 9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
use super::Kernel;
use crate::{KernelAdd, KernelError, KernelMul};
use rayon::prelude::*;
use std::{ops::Add, ops::Mul};

fn weighted_norm_pow(params: &[f64], x: &Vec<f64>, xprime: &Vec<f64>) -> f64 {
  params
    .par_iter()
    .zip(x.par_iter())
    .zip(xprime.par_iter())
    .map(|((relevance, xi), xprimei)| relevance * (xi - xprimei).powi(2))
    .sum()
}

#[derive(Clone, Debug)]
pub struct ARD(pub usize);

impl Kernel<Vec<f64>> for ARD {
  fn params_len(&self) -> usize {
    self.0
  }

  fn value(&self, params: &[f64], x: &Vec<f64>, xprime: &Vec<f64>) -> Result<f64, KernelError> {
    if params.len() != self.0 {
      return Err(KernelError::ParametersLengthMismatch.into());
    }
    if x.len() != self.0 || xprime.len() != self.0 {
      return Err(KernelError::InvalidArgument.into());
    }

    let fx = (-weighted_norm_pow(&params, x, xprime)).exp();

    Ok(fx)
  }

  fn value_with_grad(
    &self,
    params: &[f64],
    x: &Vec<f64>,
    xprime: &Vec<f64>,
  ) -> Result<(f64, Vec<f64>), KernelError> {
    let fx = self.value(params, x, xprime)?;

    let gfx = x
      .par_iter()
      .zip(xprime.par_iter())
      .map(|(&xi, &xprimei)| -(xi - xprimei).powi(2))
      .collect::<Vec<_>>();

    Ok((fx, gfx))
  }
}

impl<R> Add<R> for ARD
where
  R: Kernel<Vec<f64>>,
{
  type Output = KernelAdd<Self, R, Vec<f64>>;

  fn add(self, rhs: R) -> Self::Output {
    Self::Output::new(self, rhs)
  }
}

impl<R> Mul<R> for ARD
where
  R: Kernel<Vec<f64>>,
{
  type Output = KernelMul<Self, R, Vec<f64>>;

  fn mul(self, rhs: R) -> Self::Output {
    Self::Output::new(self, rhs)
  }
}