1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
use super::Device;
use crate::{shapes::*, tensor::*};
pub fn softmax<Ax: Axes, S: Shape, E: Dtype, D: Device<E>, T: Tape<E, D>>(
t: Tensor<S, E, D, T>,
) -> Tensor<S, E, D, T>
where
S: ReduceShape<Ax>,
{
t.softmax::<Ax>()
}
impl<S: Shape, E: Dtype, D: Device<E>, T: Tape<E, D>> Tensor<S, E, D, T> {
pub fn softmax<Ax: Axes>(self) -> Self
where
S: ReduceShape<Ax>,
{
self.try_softmax::<Ax>().unwrap()
}
pub fn try_softmax<Ax: Axes>(self) -> Result<Self, D::Err>
where
S: ReduceShape<Ax>,
{
self.try_log_softmax::<Ax>()?.try_exp()
}
}
#[cfg(test)]
mod tests {
use crate::{shapes::*, tensor::*, tensor_ops::*, tests::*};
#[test]
fn test_softmax_1d() {
let dev: TestDevice = Default::default();
let a: Tensor<_, TestDtype, _> = dev.tensor([-2.0, -1.0, 0.0, 1.0, 2.0]);
let r = a.leaky_trace().softmax();
assert_close(
&r.array(),
&[0.011656232, 0.031684924, 0.086128555, 0.23412168, 0.6364087],
);
let l = r * dev.tensor([0.0, 0.0, 1.0, 0.0, 0.0]);
assert_close(&l.array(), &[0.0, 0.0, 0.086128555, 0.0, 0.0]);
let g = l.mean().backward();
assert_close(
&g.get(&a).array(),
&[
-0.00020078686,
-0.00054579525,
0.015742086,
-0.0040329117,
-0.010962591,
],
);
}
#[test]
fn test_softmax_2d() {
let dev: TestDevice = Default::default();
let a: Tensor<_, TestDtype, _> = dev.tensor([[-2.0, -1.0, 0.0], [1.0, 4.0, 7.0]]);
let r = a.leaky_trace().softmax::<Axis<1>>();
assert_close(
&r.array(),
&[
[0.09003058, 0.24472849, 0.66524094],
[0.002355633, 0.047314156, 0.9503302],
],
);
let l = r * dev.tensor([[1.0, 0.0, 0.0], [0.0, 1.0, 0.0]]);
assert_close(
&l.array(),
&[[0.09003058, 0.0, 0.0], [0.0, 0.047314156, 0.0]],
);
let g = l.mean().backward();
assert_close(
&g.get(&a).array(),
&[
[0.01365418, -0.0036721744, -0.009982005],
[-1.85758e-5, 0.0075125876, -0.0074940124],
],
);
}
#[test]
fn test_softmax_2d_0th_axis() {
let dev: TestDevice = Default::default();
let a: Tensor<_, TestDtype, _> = dev.tensor([[-2.0, -1.0, 0.0], [1.0, 4.0, 7.0]]);
let r = a.leaky_trace().softmax::<Axis<0>>();
assert_close(
&r.array(),
&[
[0.047425874, 0.0066928514, 0.0009110514],
[0.95257413, 0.9933072, 0.9990892],
],
);
let l = r * dev.tensor([[1.0, 0.0, 0.0], [0.0, 1.0, 0.0]]);
assert_close(
&l.array(),
&[[0.047425874, 0.0, 0.0], [0.0, 0.9933072, 0.0]],
);
let g = l.mean().backward();
assert_close(
&g.get(&a).array(),
&[
[0.0075294436, -0.0011080095, 0.0],
[-0.0075294436, 0.0011080056, 0.0],
],
);
}
#[test]
fn test_softmax_3d_to_1d_12() {
let dev: TestDevice = Default::default();
let t: Tensor<Rank3<2, 3, 4>, TestDtype, _> = dev.sample_normal();
let r = t.leaky_trace().softmax::<Axes2<1, 2>>();
#[rustfmt::skip]
assert_close(
&r.array(),
&[
[[0.08535644, 0.0987266, 0.00366116, 0.04927256], [0.01169326, 0.1515922, 0.00951258, 0.07721686], [0.0776206, 0.23813945, 0.19471556, 0.00249278]],
[[0.01881982, 0.25171953, 0.02559674, 0.03725754], [0.04064152, 0.314442, 0.02427996, 0.04708378], [0.02791536, 0.14462142, 0.02221143, 0.04541067]],
],
);
}
}