neuronika 0.2.0

Tensors and dynamic neural networks.
use super::{
    assert_almost_equals, new_backward_input, new_input, new_tensor, Backward, Data, Forward,
    Gradient, MultiStack, MultiStackBackward, Overwrite, Tensor,
};

mod forward {
    use super::{assert_almost_equals, new_input, new_tensor, Data, Forward, MultiStack, Tensor};

    #[test]
    fn creation() {
        let first = new_input((3, 3), vec![-4., -3., -2., -1., 0., 1., 2., 3., 4.]);
        let second = new_input((3, 3), vec![0.; 9]);
        let node = MultiStack::new(vec![first, second], 0, new_tensor((2, 3, 3), vec![0.; 18]));

        assert_eq!(*node.data(), Tensor::from_elem((2, 3, 3), 0.));
        assert_eq!(*node.data_mut(), Tensor::from_elem((2, 3, 3), 0.));
        assert!(!node.was_computed());
    }

    #[test]
    fn computation_was_computed_transition() {
        let first = new_input((3, 3), vec![-4., -3., -2., -1., 0., 1., 2., 3., 4.]);
        let second = new_input((3, 3), vec![0.; 9]);
        let node = MultiStack::new(vec![first, second], 0, new_tensor((2, 3, 3), vec![0.; 18]));

        node.forward();
        assert!(node.was_computed());

        node.forward();
        assert!(node.was_computed());

        node.reset_computation();
        assert!(!node.was_computed());

        node.reset_computation();
        assert!(!node.was_computed());
    }

    #[test]
    fn forward() {
        let first = new_input((3, 3), vec![-4., -3., -2., -1., 0., 1., 2., 3., 4.]);
        let second = new_input((3, 3), vec![0.; 9]);
        let node = MultiStack::new(
            vec![first.clone(), second],
            0,
            new_tensor((2, 3, 3), vec![0.; 18]),
        );

        // ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ First Evaluation ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
        node.forward();
        assert_almost_equals(
            &*node.data(),
            &new_tensor(
                (2, 3, 3),
                vec![
                    -4., -3., -2., -1., 0., 1., 2., 3., 4., 0., 0., 0., 0., 0., 0., 0., 0., 0.,
                ],
            ),
        );

        // ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ No Second Evaluation ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
        {
            let mut data = first.data_mut();
            *data = &*data + &Tensor::from_elem(1, 1.);
        }
        assert_almost_equals(
            &*first.data(),
            &new_tensor((3, 3), vec![-3., -2., -1., 0., 1., 2., 3., 4., 5.]),
        );

        node.forward();
        assert_almost_equals(
            &*node.data(),
            &new_tensor(
                (2, 3, 3),
                vec![
                    -4., -3., -2., -1., 0., 1., 2., 3., 4., 0., 0., 0., 0., 0., 0., 0., 0., 0.,
                ],
            ),
        );

        // ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ Second Evaluation ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
        node.reset_computation();
        node.forward();
        assert_almost_equals(
            &*node.data(),
            &new_tensor(
                (2, 3, 3),
                vec![
                    -3., -2., -1., 0., 1., 2., 3., 4., 5., 0., 0., 0., 0., 0., 0., 0., 0., 0.,
                ],
            ),
        );
    }

    #[test]
    fn debug() {
        let first = new_input(3, vec![0., 1., 2.]);
        let second = new_input(3, vec![0.; 3]);
        let node = MultiStack::new(vec![first, second], 0, new_tensor((2, 3), vec![0.; 6]));

        let output = "MultiStack { data: [[0.0, 0.0, 0.0],\n [0.0, 0.0, 0.0]], shape=[2, 3], strides=[3, 1], layout=Cc (0x5), const ndim=2, axis: 0, operands: 2, computed: false }";

        assert_eq!(output, format!("{:?}", node));
    }

    #[test]
    fn display() {
        let first = new_input(3, vec![0., 1., 2.]);
        let second = new_input(3, vec![0.; 3]);
        let node = MultiStack::new(vec![first, second], 0, new_tensor((2, 3), vec![0.; 6]));

        assert_eq!(format!("{}", node.data()), format!("{}", node));
    }
}

mod backward {
    use super::{
        assert_almost_equals, new_backward_input, new_tensor, Backward, Gradient,
        MultiStackBackward, Overwrite, Tensor,
    };

    #[test]
    fn creation() {
        let node = MultiStackBackward::new(
            vec![
                new_backward_input((4, 3), vec![0.; 12]),
                new_backward_input((4, 3), vec![0.; 12]),
            ],
            0,
            ndarray::Dim([2, 4, 3]),
        );

        assert_eq!(*node.gradient(), Tensor::from_elem((2, 4, 3), 0.));
        assert_eq!(*node.gradient_mut(), Tensor::from_elem((2, 4, 3), 0.));
        assert!(node.can_overwrite());
    }

    #[test]
    fn computation_state_transition() {
        let first = new_backward_input((4, 3), vec![0.; 12]);
        let second = new_backward_input((4, 3), vec![0.; 12]);
        let node = MultiStackBackward::new(
            vec![first.clone(), second.clone()],
            0,
            ndarray::Dim([2, 4, 3]),
        );

        node.backward();
        assert!(node.can_overwrite());
        assert!(!first.can_overwrite());
        assert!(!second.can_overwrite());

        node.backward();
        assert!(node.can_overwrite());
        assert!(!first.can_overwrite());
        assert!(!second.can_overwrite());

        first.set_overwrite(true);
        assert!(node.can_overwrite());
        assert!(first.can_overwrite());
        assert!(!second.can_overwrite());

        first.set_overwrite(true);
        assert!(node.can_overwrite());
        assert!(first.can_overwrite());
        assert!(!second.can_overwrite());

        second.set_overwrite(true);
        assert!(node.can_overwrite());
        assert!(first.can_overwrite());
        assert!(second.can_overwrite());

        second.set_overwrite(true);
        assert!(node.can_overwrite());
        assert!(first.can_overwrite());
        assert!(second.can_overwrite());

        node.set_overwrite(false);
        assert!(!node.can_overwrite());
        assert!(first.can_overwrite());
        assert!(second.can_overwrite());

        node.set_overwrite(false);
        assert!(!node.can_overwrite());
        assert!(first.can_overwrite());
        assert!(second.can_overwrite());

        node.backward();
        assert!(!node.can_overwrite());
        assert!(!first.can_overwrite());
        assert!(!second.can_overwrite());

        node.backward();
        assert!(!node.can_overwrite());
        assert!(!first.can_overwrite());
        assert!(!second.can_overwrite());
    }

    #[test]
    fn backward() {
        let first = new_backward_input((4, 3), vec![0.; 12]);
        let second = new_backward_input((4, 3), vec![0.; 12]);
        let node = MultiStackBackward::new(
            vec![first.clone(), second.clone()],
            0,
            ndarray::Dim([2, 4, 3]),
        );

        // ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ Seed Gradient ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
        *node.gradient_mut() = new_tensor((2, 4, 3), vec![1.; 24]);
        assert_almost_equals(&*node.gradient(), &new_tensor((2, 4, 3), vec![1.; 24]));

        // ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ First Evaluation ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
        node.backward();
        assert_almost_equals(&*first.gradient(), &new_tensor((4, 3), vec![1.; 12]));
        assert_almost_equals(&*second.gradient(), &new_tensor((4, 3), vec![1.; 12]));

        // ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ Second Evaluation ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
        node.backward();
        assert_almost_equals(&*first.gradient(), &new_tensor((4, 3), vec![2.; 12]));
        assert_almost_equals(&*second.gradient(), &new_tensor((4, 3), vec![2.; 12]));

        // ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ Third Evaluation ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
        first.set_overwrite(true);
        second.set_overwrite(true);
        node.backward();
        assert_almost_equals(&*first.gradient(), &new_tensor((4, 3), vec![1.; 12]));
        assert_almost_equals(&*second.gradient(), &new_tensor((4, 3), vec![1.; 12]));
    }

    #[test]
    fn no_grad() {
        // MultiStackBackward
        let node = MultiStackBackward::new(
            vec![
                new_backward_input((3, 3), vec![0.; 9]),
                new_backward_input((3, 3), vec![0.; 9]),
            ],
            0,
            ndarray::Dim([2, 3, 3]),
        );

        node.no_grad();
        assert!(node.gradient.borrow().is_none());

        node.with_grad();
        assert_eq!(&*node.gradient(), Tensor::zeros(node.shape));
    }

    #[test]
    fn debug() {
        let first = new_backward_input(3, vec![0.; 3]);
        let second = new_backward_input(3, vec![0.; 3]);
        let node =
            MultiStackBackward::new(vec![first.clone(), second.clone()], 0, ndarray::Dim([2, 3]));

        let output = "MultiStackBackward { gradient: Some([[0.0, 0.0, 0.0],\n [0.0, 0.0, 0.0]], shape=[2, 3], strides=[3, 1], layout=Cc (0x5), const ndim=2), operands: 2, axis: 0, overwrite: Cell { value: true } }";

        assert_eq!(output, format!("{:?}", node));
    }

    #[test]
    fn display() {
        let first = new_backward_input(3, vec![0.; 3]);
        let second = new_backward_input(3, vec![0.; 3]);
        let node =
            MultiStackBackward::new(vec![first.clone(), second.clone()], 0, ndarray::Dim([2, 3]));

        assert_eq!(format!("{}", node.gradient()), format!("{}", node));
    }
}