Expand description
BERT (Bidirectional Encoder Representations from Transformers)
Implements BERT as described in “BERT: Pre-training of Deep Bidirectional Transformers”
- Token embeddings
- Segment embeddings
- Position embeddings
- Transformer encoder layers
- Masked language modeling head
- Next sentence prediction head
Structs§
- Bert
Config - BERT configuration
- Bert
Embeddings - BERT embeddings layer
- Bert
ForMaskedLM - BERT for Masked Language Modeling
- Bert
ForSequence Classification - BERT for Sequence Classification
- Bert
ForToken Classification - BERT for Token Classification (NER, POS tagging)
- Bert
Model - BERT model
- Bert
Output - BERT output
- Bert
Pooler - BERT pooler (for classification tasks)