Expand description
FastGRNN training pipeline with knowledge distillation
This module provides a complete training infrastructure for the FastGRNN model:
- Adam optimizer implementation
- Binary Cross-Entropy loss with gradient computation
- Backpropagation Through Time (BPTT)
- Mini-batch training with validation split
- Early stopping and learning rate scheduling
- Knowledge distillation from teacher models
- Progress reporting and metrics tracking
Structs§
- Batch
Iterator - Batch iterator for training
- Trainer
- FastGRNN trainer
- Training
Config - Training hyperparameters
- Training
Dataset - Training dataset with features and labels
- Training
Metrics - Training metrics
Functions§
- generate_
teacher_ predictions - Generate teacher predictions for knowledge distillation
- temperature_
softmax - Temperature-scaled softmax for knowledge distillation with numerical stability