use burn_core as burn;
use burn::config::Config;
use burn::module::Param;
use burn::module::{Content, DisplaySettings, Initializer, Module, ModuleDisplay};
use burn::tensor::module::linear;
use burn::tensor::{Tensor, backend::Backend};
#[derive(Config, Debug)]
pub struct LinearConfig {
pub d_input: usize,
pub d_output: usize,
#[config(default = true)]
pub bias: bool,
#[config(
default = "Initializer::KaimingUniform{gain:1.0/num_traits::Float::sqrt(3.0), fan_out_only:false}"
)]
pub initializer: Initializer,
#[config(default = "LinearLayout::Row")]
pub layout: LinearLayout,
}
#[derive(Config, Debug, Copy)]
pub enum LinearLayout {
Row,
Col,
}
#[derive(Module, Debug)]
#[module(custom_display)]
pub struct Linear<B: Backend> {
pub weight: Param<Tensor<B, 2>>,
pub bias: Option<Param<Tensor<B, 1>>>,
}
impl LinearConfig {
pub fn init<B: Backend>(&self, device: &B::Device) -> Linear<B> {
let weight = match self.layout {
LinearLayout::Row => {
let shape = [self.d_input, self.d_output];
self.initializer
.init_with(shape, Some(self.d_input), Some(self.d_output), device)
}
LinearLayout::Col => {
let shape = [self.d_output, self.d_input];
self.initializer
.init_with(shape, Some(self.d_output), Some(self.d_input), device)
.save_mapper(move |tensor| {
B::sync(&tensor.device()).unwrap();
let tensor = tensor.transpose();
B::sync(&tensor.device()).unwrap();
tensor
})
.load_mapper(move |tensor| {
B::sync(&tensor.device()).unwrap();
let tensor = tensor.transpose();
B::sync(&tensor.device()).unwrap();
tensor
})
.init_mapper(|tensor| {
B::sync(&tensor.device()).unwrap();
let tensor = tensor.transpose();
B::sync(&tensor.device()).unwrap();
tensor
})
}
};
let bias = if self.bias {
Some(self.initializer.init_with(
[self.d_output],
Some(self.d_input),
Some(self.d_output),
device,
))
} else {
None
};
Linear { weight, bias }
}
}
impl<B: Backend> Linear<B> {
pub fn forward<const D: usize>(&self, input: Tensor<B, D>) -> Tensor<B, D> {
linear(
input,
self.weight.val(),
self.bias.as_ref().map(|b| b.val()),
)
}
}
impl<B: Backend> ModuleDisplay for Linear<B> {
fn custom_settings(&self) -> Option<DisplaySettings> {
DisplaySettings::new()
.with_new_line_after_attribute(false)
.optional()
}
fn custom_content(&self, content: Content) -> Option<Content> {
let [d_input, d_output] = self.weight.shape().dims();
content
.add("d_input", &d_input)
.add("d_output", &d_output)
.add("bias", &self.bias.is_some())
.optional()
}
}
#[cfg(test)]
mod tests {
use super::*;
use crate::TestBackend;
use burn::module::ParamId;
use burn::record::{BinBytesRecorder, FullPrecisionSettings, Recorder};
use burn::tensor::ElementConversion;
use burn::tensor::{Shape, TensorData};
use burn::tensor::{Tolerance, ops::FloatElem};
type FT = FloatElem<TestBackend>;
#[test]
fn initializer_default() {
let device = Default::default();
TestBackend::seed(&device, 0);
let config = LinearConfig::new(5, 5);
let k = (1.0 / config.d_input as f64).sqrt().elem::<FT>();
let linear = config.init::<TestBackend>(&device);
assert_eq!(
config.initializer,
Initializer::KaimingUniform {
gain: 1.0 / 3.0f64.sqrt(),
fan_out_only: false
}
);
linear.weight.to_data().assert_within_range(-k..k);
}
#[test]
fn initializer_zeros() {
let device = Default::default();
TestBackend::seed(&device, 0);
let config = LinearConfig::new(5, 5).with_initializer(Initializer::Zeros);
let linear = config.init::<TestBackend>(&device);
assert_eq!(config.initializer, Initializer::Zeros);
linear.weight.to_data().assert_approx_eq::<FT>(
&TensorData::zeros::<f32, _>(linear.weight.shape()),
Tolerance::default(),
);
}
#[test]
fn test_linear_forward_no_bias() {
let device = Default::default();
TestBackend::seed(&device, 0);
let value = 2.;
let config = LinearConfig::new(2, 3)
.with_initializer(Initializer::Constant { value })
.with_bias(false);
let linear = config.init::<TestBackend>(&device);
let input = Tensor::<TestBackend, 2>::ones(Shape::new([1, 2]), &device);
let result = linear.forward(input);
let expected_result = Tensor::<TestBackend, 2>::from_data([[4., 4., 4.]], &device);
assert_eq!(result.into_data(), expected_result.into_data());
}
#[test]
fn test_linear_forward_with_bias() {
let device = Default::default();
TestBackend::seed(&device, 0);
let device = Default::default();
let value = 2.;
let config = LinearConfig::new(2, 3).with_initializer(Initializer::Constant { value });
let linear = config.init::<TestBackend>(&device);
let input = Tensor::<TestBackend, 2>::ones(Shape::new([1, 2]), &device);
let result = linear.forward(input);
let expected_result = Tensor::<TestBackend, 2>::from_data([[6., 6., 6.]], &device);
assert_eq!(result.into_data(), expected_result.into_data());
}
#[test]
fn test_linear_1d() {
let device = Default::default();
TestBackend::seed(&device, 0);
let device = Default::default();
let value = 2.;
let config = LinearConfig::new(2, 3).with_initializer(Initializer::Constant { value });
let linear = config.init::<TestBackend>(&device);
let input_1d = Tensor::<TestBackend, 1>::ones(Shape::new([2]), &device);
let input_2d = Tensor::<TestBackend, 2>::ones(Shape::new([1, 2]), &device);
let result_1d = linear.forward(input_1d).unsqueeze::<2>();
let result_2d = linear.forward(input_2d);
assert_eq!(result_1d.into_data(), result_2d.into_data());
}
#[test]
fn display() {
let config = LinearConfig::new(3, 5);
let linear = config.init::<TestBackend>(&Default::default());
assert_eq!(
alloc::format!("{linear}"),
"Linear {d_input: 3, d_output: 5, bias: true, params: 20}"
);
}
#[test]
fn layout() {
let device = Default::default();
let config = LinearConfig::new(6, 12).with_layout(LinearLayout::Col);
let linear = config.init::<TestBackend>(&device);
assert_eq!(linear.weight.dims(), [6, 12], "Shape is as configured");
let recorder = BinBytesRecorder::<FullPrecisionSettings>::new();
let record = linear.into_record();
let data = recorder.record(record, ()).unwrap();
let record = recorder.load(data.clone(), &device).unwrap();
let config = LinearConfig::new(12, 6).with_layout(LinearLayout::Row);
let linear_row = config.init::<TestBackend>(&device).load_record(record);
assert_eq!(
linear_row.weight.dims(),
[12, 6],
"Shape should be transposed"
);
let record = recorder.load(data.clone(), &device).unwrap();
let config = LinearConfig::new(6, 12).with_layout(LinearLayout::Col);
let linear_col = config.init::<TestBackend>(&device).load_record(record);
assert_eq!(
linear_col.weight.dims(),
[6, 12],
"Shape should be as configured"
);
let record = linear_col.into_record();
let data = recorder.record(record, ()).unwrap();
let record = recorder.load(data, &device).unwrap();
let config = LinearConfig::new(6, 12).with_layout(LinearLayout::Col);
let linear_col = config.init::<TestBackend>(&device).load_record(record);
assert_eq!(
linear_col.weight.dims(),
[6, 12],
"Shape should be as configured"
);
}
#[test]
fn col_row_same_result() {
let device = Default::default();
let config_col = LinearConfig::new(6, 12).with_layout(LinearLayout::Col);
let linear_col = config_col.init::<TestBackend>(&device);
let signal = Tensor::<_, 2>::random([8, 6], burn::tensor::Distribution::Default, &device);
let value = linear_col.forward(signal.clone());
let data_1 = value.into_data();
let weights = linear_col.weight.val().into_data();
let weights = Tensor::from_data(weights, &device);
let linear = Linear {
weight: Param::initialized(ParamId::new(), weights),
bias: linear_col
.bias
.map(|b| Param::initialized(ParamId::new(), b.val())),
};
let value = linear.forward(signal);
let data_2 = value.into_data();
data_1.assert_approx_eq::<f32>(&data_2, Default::default());
}
}