syntaxdot-transformers 0.5.0

Transformer architectures, such as BERT
Documentation
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
//! SqueezeBERT (Iandola et al., 2020)
//!
//! SqueezeBERT follows the same architecture as BERT, but replaces most
//! matrix multiplications by grouped convolutions. This reduces the
//! number of parameters and speeds up inference.

mod config;
pub use config::SqueezeBertConfig;

mod embeddings;

mod encoder;
pub use encoder::SqueezeBertEncoder;

mod layer;
pub(crate) use layer::SqueezeBertLayer;