axonml-optim 0.1.0

Optimizers and learning rate schedulers for the Axonml ML framework
Documentation
# axonml-optim

Optimizers and learning rate schedulers for [Axonml](https://github.com/AutomataNexus/AxonML).

## Optimizers

- **SGD** - With momentum, Nesterov, weight decay
- **Adam** - Adaptive learning rates
- **AdamW** - Decoupled weight decay
- **RMSprop** - Moving average of squared gradients

## LR Schedulers

- StepLR, MultiStepLR, ExponentialLR
- CosineAnnealingLR, OneCycleLR, WarmupLR

## Usage

```rust
use axonml_optim::{Adam, Optimizer};

let mut optimizer = Adam::new(model.parameters(), 0.001);
optimizer.zero_grad();
loss.backward();
optimizer.step();
```

## License

MIT OR Apache-2.0