use burn_core as burn;
use burn::config::Config;
use burn::tensor::backend::Backend;
use super::{LrScheduler, String};
use crate::LearningRate;
#[derive(Config, Debug)]
pub struct NoamLrSchedulerConfig {
factor: f64,
#[config(default = 4000)]
warmup_steps: usize,
#[config(default = 512)]
model_size: usize,
}
#[derive(Clone, Debug)]
pub struct NoamLrScheduler {
warmup_steps: f64,
embedding_size: f64,
factor: f64,
step: f64,
}
impl NoamLrSchedulerConfig {
pub fn init(&self) -> Result<NoamLrScheduler, String> {
if self.warmup_steps == 0 {
return Err(
"Number of steps before exponential decay starts must be greater than 0".into(),
);
}
if self.model_size == 0 {
return Err("Model size must be greater than 0".into());
}
Ok(NoamLrScheduler {
warmup_steps: self.warmup_steps as f64,
embedding_size: self.model_size as f64,
factor: self.factor,
step: 0.0,
})
}
}
impl LrScheduler for NoamLrScheduler {
type Record<B: Backend> = usize;
fn step(&mut self) -> LearningRate {
self.step += 1.0;
let arg1 = self.step.powf(-0.5);
let arg2 = self.step * self.warmup_steps.powf(-1.5);
self.factor * self.embedding_size.powf(-0.5) * f64::min(arg1, arg2)
}
fn to_record<B: Backend>(&self) -> Self::Record<B> {
self.step as usize
}
fn load_record<B: Backend>(mut self, record: Self::Record<B>) -> Self {
self.step = record as f64;
self
}
}
#[cfg(test)]
mod tests {
use super::*;
#[test]
fn test_config_warmup_steps_invalid() {
let r = NoamLrSchedulerConfig::new(0.1).with_warmup_steps(0).init();
assert!(r.is_err(), "Should return an error");
}
#[test]
fn test_config_warmup_steps_valid() {
let r = NoamLrSchedulerConfig::new(0.1).with_warmup_steps(1).init();
assert!(r.is_ok(), "Should return a success value");
}
#[test]
fn test_config_model_size_invalid() {
let r = NoamLrSchedulerConfig::new(0.1).with_model_size(0).init();
assert!(r.is_err(), "Should return an error");
}
#[test]
fn test_config_model_size_valid() {
let r = NoamLrSchedulerConfig::new(0.1).with_model_size(1).init();
assert!(r.is_ok(), "Should return a success value");
}
#[test]
fn test_function_increase_and_decrease() {
let warmup_steps = 100;
let mut scheduler = NoamLrSchedulerConfig::new(10.0)
.with_warmup_steps(warmup_steps)
.init()
.unwrap();
let mut lr_current = 0.0;
for _ in 0..warmup_steps {
let lr = scheduler.step();
assert!(
lr > lr_current,
"Learning rate should increase before the warmup_steps is reached."
);
lr_current = lr;
}
for _ in 0..warmup_steps {
let lr = scheduler.step();
assert!(
lr < lr_current,
"Learning rate should decrease after the warmup_steps is reached."
);
lr_current = lr;
}
}
}