Expand description
Transformer models.
Modules§
- albert
- ALBERT (Lan et al., 2020)
- bert
- BERT (Devlin et al., 2018)
- roberta
- RoBERTa (Liu et al., 2018) and XLM-RoBERTa (Conneau et al., 2019).
- sinusoidal
- Word embeddings with sinusoidal position embeddings.
- squeeze_
albert - SqueezeBERT (Iandola et al., 2020) + ALBERT (Lan et al., 2020)
- squeeze_
bert - SqueezeBERT (Iandola et al., 2020)
Structs§
- Hidden
Layer - Hidden layer output and attention.
Enums§
- Layer
Output - Output of a BERT layer.
Traits§
- Encoder
- Encoder networks.