use super::{
assert_almost_equals, new_backward_input, new_input, new_tensor, Backward, Data, Forward,
Gradient, MatrixMatrixMul, MatrixMatrixMulBackward, MatrixMatrixMulBackwardLeft,
MatrixMatrixMulBackwardRight, Overwrite, Tensor,
};
#[cfg(feature = "blas")]
extern crate blas_src;
mod forward {
use super::{
assert_almost_equals, new_input, new_tensor, Data, Forward, MatrixMatrixMul, Tensor,
};
#[test]
fn creation() {
let left = new_input((3, 3), vec![1., 2., 3., 4., 5., 6., 7., 8., 9.]);
let right = new_input((3, 3), vec![1.; 9]);
let node = MatrixMatrixMul::new(left, right);
assert_eq!(*node.data(), Tensor::from_elem((3, 3), 0.));
assert_eq!(*node.data_mut(), Tensor::from_elem((3, 3), 0.));
assert!(!node.was_computed());
}
#[test]
fn computation_was_computed_transition() {
let left = new_input((3, 3), vec![1., 2., 3., 4., 5., 6., 7., 8., 9.]);
let right = new_input((3, 3), vec![1.; 9]);
let node = MatrixMatrixMul::new(left, right);
node.forward();
assert!(node.was_computed());
node.forward();
assert!(node.was_computed());
node.reset_computation();
assert!(!node.was_computed());
node.reset_computation();
assert!(!node.was_computed());
}
#[test]
fn forward() {
let left = new_input((3, 3), vec![1., 2., 3., 4., 5., 6., 7., 8., 9.]);
let right = new_input((3, 3), vec![1.; 9]);
let node = MatrixMatrixMul::new(left, right.clone());
node.forward();
assert_almost_equals(
&*node.data(),
&new_tensor((3, 3), vec![6., 6., 6., 15., 15., 15., 24., 24., 24.]),
);
*right.data_mut() = new_tensor((3, 3), vec![-2.; 9]);
assert_almost_equals(&*right.data(), &new_tensor((3, 3), vec![-2.; 9]));
node.forward();
assert_almost_equals(
&*node.data(),
&new_tensor((3, 3), vec![6., 6., 6., 15., 15., 15., 24., 24., 24.]),
);
node.reset_computation();
node.forward();
assert_almost_equals(
&*node.data(),
&new_tensor(
(3, 3),
vec![-12., -12., -12., -30., -30., -30., -48., -48., -48.],
),
);
}
#[test]
fn debug() {
let left = new_input((3, 3), vec![1., 2., 3., 4., 5., 6., 7., 8., 9.]);
let right = new_input((3, 3), vec![1.; 9]);
let node = MatrixMatrixMul::new(left, right);
let output = "MatrixMatrixMul { data: [[0.0, 0.0, 0.0],\n [0.0, 0.0, 0.0],\n [0.0, 0.0, 0.0]], shape=[3, 3], strides=[3, 1], layout=Cc (0x5), const ndim=2, computed: false }";
assert_eq!(output, format!("{:?}", node));
}
#[test]
fn display() {
let left = new_input((3, 3), vec![1., 2., 3., 4., 5., 6., 7., 8., 9.]);
let right = new_input((3, 3), vec![1.; 9]);
let node = MatrixMatrixMul::new(left, right);
assert_eq!(format!("{}", node.data()), format!("{}", node));
}
}
mod backward {
use super::{
assert_almost_equals, new_backward_input, new_input, new_tensor, Backward, Gradient,
MatrixMatrixMulBackward, MatrixMatrixMulBackwardLeft, MatrixMatrixMulBackwardRight,
Overwrite, Tensor,
};
#[test]
fn creation() {
let node = MatrixMatrixMulBackward::new(
new_input((3, 3), vec![1., 2., 3., 4., 5., 6., 7., 8., 9.]),
new_backward_input((3, 3), vec![0.; 9]),
new_input((3, 3), vec![10., 11., 12., 13., 14., 15., 16., 17., 18.]),
new_backward_input((3, 3), vec![0.; 9]),
);
assert_eq!(*node.gradient(), Tensor::from_elem((3, 3), 0.));
assert_eq!(*node.gradient_mut(), Tensor::from_elem((3, 3), 0.));
assert!(node.can_overwrite());
}
#[test]
fn computation_state_transition() {
let lhs = new_backward_input((3, 3), vec![0.; 9]);
let rhs = new_backward_input((3, 3), vec![0.; 9]);
let node = MatrixMatrixMulBackward::new(
new_input((3, 3), vec![1., 2., 3., 4., 5., 6., 7., 8., 9.]),
lhs.clone(),
new_input((3, 3), vec![10., 11., 12., 13., 14., 15., 16., 17., 18.]),
rhs.clone(),
);
node.backward();
assert!(node.can_overwrite());
assert!(!lhs.can_overwrite());
assert!(!rhs.can_overwrite());
node.backward();
assert!(node.can_overwrite());
assert!(!lhs.can_overwrite());
assert!(!rhs.can_overwrite());
lhs.set_overwrite(true);
assert!(node.can_overwrite());
assert!(lhs.can_overwrite());
assert!(!rhs.can_overwrite());
lhs.set_overwrite(true);
assert!(node.can_overwrite());
assert!(lhs.can_overwrite());
assert!(!rhs.can_overwrite());
rhs.set_overwrite(true);
assert!(node.can_overwrite());
assert!(lhs.can_overwrite());
assert!(rhs.can_overwrite());
rhs.set_overwrite(true);
assert!(node.can_overwrite());
assert!(lhs.can_overwrite());
assert!(rhs.can_overwrite());
node.set_overwrite(false);
assert!(!node.can_overwrite());
assert!(lhs.can_overwrite());
assert!(rhs.can_overwrite());
node.set_overwrite(false);
assert!(!node.can_overwrite());
assert!(lhs.can_overwrite());
assert!(rhs.can_overwrite());
node.backward();
assert!(!node.can_overwrite());
assert!(!lhs.can_overwrite());
assert!(!rhs.can_overwrite());
node.backward();
assert!(!node.can_overwrite());
assert!(!lhs.can_overwrite());
assert!(!rhs.can_overwrite());
}
#[test]
fn backward() {
let lhs = new_backward_input((3, 3), vec![0.; 9]);
let rhs = new_backward_input((3, 3), vec![0.; 9]);
let node = MatrixMatrixMulBackward::new(
new_input((3, 3), vec![1., 2., 3., 4., 5., 6., 7., 8., 9.]),
lhs.clone(),
new_input((3, 3), vec![10., 11., 12., 13., 14., 15., 16., 17., 18.]),
rhs.clone(),
);
*node.gradient_mut() = new_tensor((3, 3), vec![1.; 9]);
assert_almost_equals(&*node.gradient(), &new_tensor((3, 3), vec![1.; 9]));
node.backward();
assert_almost_equals(
&*lhs.gradient(),
&new_tensor((3, 3), vec![33., 42., 51., 33., 42., 51., 33., 42., 51.]),
);
assert_almost_equals(
&*rhs.gradient(),
&new_tensor((3, 3), vec![12., 12., 12., 15., 15., 15., 18., 18., 18.]),
);
node.backward();
assert_almost_equals(
&*lhs.gradient(),
&new_tensor((3, 3), vec![66., 84., 102., 66., 84., 102., 66., 84., 102.]),
);
assert_almost_equals(
&*rhs.gradient(),
&new_tensor((3, 3), vec![24., 24., 24., 30., 30., 30., 36., 36., 36.]),
);
lhs.set_overwrite(true);
rhs.set_overwrite(true);
node.backward();
assert_almost_equals(
&*lhs.gradient(),
&new_tensor((3, 3), vec![33., 42., 51., 33., 42., 51., 33., 42., 51.]),
);
assert_almost_equals(
&*rhs.gradient(),
&new_tensor((3, 3), vec![12., 12., 12., 15., 15., 15., 18., 18., 18.]),
);
}
#[test]
fn debug() {
let lhs = new_backward_input((3, 3), vec![0.; 9]);
let rhs = new_backward_input((3, 3), vec![0.; 9]);
let node = MatrixMatrixMulBackward::new(
new_input((3, 3), vec![1., 2., 3., 4., 5., 6., 7., 8., 9.]),
lhs,
new_input((3, 3), vec![10., 11., 12., 13., 14., 15., 16., 17., 18.]),
rhs,
);
let output = "MatrixMatrixMulBackward { gradient: Some([[0.0, 0.0, 0.0],\n [0.0, 0.0, 0.0],\n [0.0, 0.0, 0.0]], shape=[3, 3], strides=[3, 1], layout=Cc (0x5), const ndim=2), overwrite: true }";
assert_eq!(output, format!("{:?}", node));
}
#[test]
fn display() {
let lhs = new_backward_input((3, 3), vec![0.; 9]);
let rhs = new_backward_input((3, 3), vec![0.; 9]);
let node = MatrixMatrixMulBackward::new(
new_input((3, 3), vec![1., 2., 3., 4., 5., 6., 7., 8., 9.]),
lhs,
new_input((3, 3), vec![10., 11., 12., 13., 14., 15., 16., 17., 18.]),
rhs,
);
assert_eq!(format!("{}", node.gradient()), format!("{}", node));
}
#[test]
fn backward_left() {
let diff = new_backward_input((3, 3), vec![0.; 9]);
let node = MatrixMatrixMulBackwardLeft::new(
diff.clone(),
new_input((3, 3), vec![10., 11., 12., 13., 14., 15., 16., 17., 18.]),
);
*node.gradient_mut() = new_tensor((3, 3), vec![1.; 9]);
assert_almost_equals(&*node.gradient(), &new_tensor((3, 3), vec![1.; 9]));
node.backward();
assert_almost_equals(
&*diff.gradient(),
&new_tensor((3, 3), vec![33., 42., 51., 33., 42., 51., 33., 42., 51.]),
);
node.backward();
assert_almost_equals(
&*diff.gradient(),
&new_tensor((3, 3), vec![66., 84., 102., 66., 84., 102., 66., 84., 102.]),
);
diff.set_overwrite(true);
node.backward();
assert_almost_equals(
&*diff.gradient(),
&new_tensor((3, 3), vec![33., 42., 51., 33., 42., 51., 33., 42., 51.]),
);
}
#[test]
fn debug_left() {
let diff = new_backward_input((3, 3), vec![0.; 9]);
let node = MatrixMatrixMulBackwardLeft::new(
diff.clone(),
new_input((3, 3), vec![10., 11., 12., 13., 14., 15., 16., 17., 18.]),
);
let output = "MatrixMatrixMulBackwardLeft { gradient: Some([[0.0, 0.0, 0.0],\n [0.0, 0.0, 0.0],\n [0.0, 0.0, 0.0]], shape=[3, 3], strides=[3, 1], layout=Cc (0x5), const ndim=2), overwrite: true }";
assert_eq!(output, format!("{:?}", node));
}
#[test]
fn display_left() {
let diff = new_backward_input((3, 3), vec![0.; 9]);
let node = MatrixMatrixMulBackwardLeft::new(
diff.clone(),
new_input((3, 3), vec![10., 11., 12., 13., 14., 15., 16., 17., 18.]),
);
assert_eq!(format!("{}", node.gradient()), format!("{}", node));
}
#[test]
fn backward_right() {
let diff = new_backward_input((3, 3), vec![0.; 9]);
let node = MatrixMatrixMulBackwardRight::new(
new_input((3, 3), vec![1., 2., 3., 4., 5., 6., 7., 8., 9.]),
diff.clone(),
);
*node.gradient_mut() = new_tensor((3, 3), vec![1.; 9]);
assert_almost_equals(&*node.gradient(), &new_tensor((3, 3), vec![1.; 9]));
node.backward();
assert_almost_equals(
&*diff.gradient(),
&new_tensor((3, 3), vec![12., 12., 12., 15., 15., 15., 18., 18., 18.]),
);
node.backward();
assert_almost_equals(
&*diff.gradient(),
&new_tensor((3, 3), vec![24., 24., 24., 30., 30., 30., 36., 36., 36.]),
);
diff.set_overwrite(true);
node.backward();
assert_almost_equals(
&*diff.gradient(),
&new_tensor((3, 3), vec![12., 12., 12., 15., 15., 15., 18., 18., 18.]),
);
}
#[test]
fn debug_right() {
let diff = new_backward_input((3, 3), vec![0.; 9]);
let node = MatrixMatrixMulBackwardRight::new(
new_input((3, 3), vec![1., 2., 3., 4., 5., 6., 7., 8., 9.]),
diff.clone(),
);
let output = "MatrixMatrixMulBackwardRight { gradient: Some([[0.0, 0.0, 0.0],\n [0.0, 0.0, 0.0],\n [0.0, 0.0, 0.0]], shape=[3, 3], strides=[3, 1], layout=Cc (0x5), const ndim=2), overwrite: true }";
assert_eq!(output, format!("{:?}", node));
}
#[test]
fn display_right() {
let diff = new_backward_input((3, 3), vec![0.; 9]);
let node = MatrixMatrixMulBackwardRight::new(
new_input((3, 3), vec![1., 2., 3., 4., 5., 6., 7., 8., 9.]),
diff.clone(),
);
assert_eq!(format!("{}", node.gradient()), format!("{}", node));
}
#[test]
fn no_grad() {
let node = MatrixMatrixMulBackward::new(
new_input((3, 3), vec![0.; 9]),
new_backward_input((3, 3), vec![0.; 9]),
new_input((3, 3), vec![0.; 9]),
new_backward_input((3, 3), vec![0.; 9]),
);
node.no_grad();
assert!(node.gradient.borrow().is_none());
node.with_grad();
assert_eq!(&*node.gradient(), Tensor::zeros(node.shape));
let node = MatrixMatrixMulBackwardLeft::new(
new_backward_input((3, 3), vec![0.; 9]),
new_input((3, 3), vec![0.; 9]),
);
node.no_grad();
assert!(node.gradient.borrow().is_none());
node.with_grad();
assert_eq!(&*node.gradient(), Tensor::zeros(node.shape));
let node = MatrixMatrixMulBackwardRight::new(
new_input((3, 3), vec![0.; 9]),
new_backward_input((3, 3), vec![0.; 9]),
);
node.no_grad();
assert!(node.gradient.borrow().is_none());
node.with_grad();
assert_eq!(&*node.gradient(), Tensor::zeros(node.shape));
}
}