use std::borrow::Borrow;
use syntaxdot_tch_ext::PathExt;
use tch::{Kind, Tensor};
use crate::cow::CowTensor;
use crate::models::bert::{BertConfig, BertEmbeddings};
use crate::module::FallibleModuleT;
use crate::TransformerError;
const PADDING_IDX: i64 = 1;
#[derive(Debug)]
pub struct RobertaEmbeddings {
inner: BertEmbeddings,
}
impl RobertaEmbeddings {
pub fn new<'a>(
vs: impl Borrow<PathExt<'a>>,
config: &BertConfig,
) -> Result<RobertaEmbeddings, TransformerError> {
Ok(RobertaEmbeddings {
inner: BertEmbeddings::new(vs, config)?,
})
}
pub fn forward(
&self,
input_ids: &Tensor,
token_type_ids: Option<&Tensor>,
position_ids: Option<&Tensor>,
train: bool,
) -> Result<Tensor, TransformerError> {
let position_ids = match position_ids {
Some(position_ids) => CowTensor::Borrowed(position_ids),
None => {
let mask = input_ids.f_ne(PADDING_IDX)?.to_kind(Kind::Int64);
let incremental_indices = mask.f_cumsum(1, Kind::Int64)?.f_mul(&mask)?;
CowTensor::Owned(incremental_indices.f_add_scalar(PADDING_IDX)?)
}
};
self.inner.forward(
input_ids,
token_type_ids,
Some(position_ids.as_ref()),
train,
)
}
}
impl FallibleModuleT for RobertaEmbeddings {
type Error = TransformerError;
fn forward_t(&self, input: &Tensor, train: bool) -> Result<Tensor, Self::Error> {
self.forward(input, None, None, train)
}
}
#[cfg(feature = "model-tests")]
#[cfg(test)]
mod tests {
use std::convert::TryInto;
use approx::assert_abs_diff_eq;
use ndarray::{array, ArrayD};
use syntaxdot_tch_ext::tensor::SumDim;
use syntaxdot_tch_ext::RootExt;
use tch::nn::VarStore;
use tch::{Device, Kind, Tensor};
use crate::activations::Activation;
use crate::models::bert::{BertConfig, BertEncoder};
use crate::models::roberta::RobertaEmbeddings;
use crate::models::Encoder;
use crate::module::FallibleModuleT;
const XLM_ROBERTA_BASE: &str = env!("XLM_ROBERTA_BASE");
fn xlm_roberta_config() -> BertConfig {
BertConfig {
attention_probs_dropout_prob: 0.1,
hidden_act: Activation::Gelu,
hidden_dropout_prob: 0.1,
hidden_size: 768,
initializer_range: 0.02,
intermediate_size: 3072,
layer_norm_eps: 1e-5,
max_position_embeddings: 514,
num_attention_heads: 12,
num_hidden_layers: 12,
type_vocab_size: 1,
vocab_size: 250002,
}
}
#[test]
fn xlm_roberta_embeddings() {
let config = xlm_roberta_config();
let mut vs = VarStore::new(Device::Cpu);
let root = vs.root_ext(|_| 0);
let embeddings = RobertaEmbeddings::new(root.sub("embeddings"), &config).unwrap();
vs.load(XLM_ROBERTA_BASE).unwrap();
let pieces = Tensor::of_slice(&[
0i64, 310, 23451, 107, 6743, 68, 62, 43789, 207126, 49004, 705, 2,
])
.reshape(&[1, 12]);
let summed_embeddings =
embeddings
.forward_t(&pieces, false)
.unwrap()
.sum_dim(-1, false, Kind::Float);
let sums: ArrayD<f32> = (&summed_embeddings).try_into().unwrap();
assert_abs_diff_eq!(
sums,
(array![[
-9.1686, -4.2982, -0.7808, -0.7097, 0.0972, -3.0785, -3.6755, -2.1465, -2.9406,
-1.0627, -6.6043, -4.8064
]])
.into_dyn(),
epsilon = 1e-4
);
}
#[test]
fn xlm_roberta_encoder() {
let config = xlm_roberta_config();
let mut vs = VarStore::new(Device::Cpu);
let root = vs.root_ext(|_| 0);
let embeddings = RobertaEmbeddings::new(root.sub("embeddings"), &config).unwrap();
let encoder = BertEncoder::new(root.sub("encoder"), &config).unwrap();
vs.load(XLM_ROBERTA_BASE).unwrap();
let pieces = Tensor::of_slice(&[
0i64, 310, 23451, 107, 6743, 68, 62, 43789, 207126, 49004, 705, 2,
])
.reshape(&[1, 12]);
let embeddings = embeddings.forward_t(&pieces, false).unwrap();
let all_hidden_states = encoder.encode(&embeddings, None, false).unwrap();
let summed_last_hidden =
all_hidden_states
.last()
.unwrap()
.output()
.sum_dim(-1, false, Kind::Float);
let sums: ArrayD<f32> = (&summed_last_hidden).try_into().unwrap();
assert_abs_diff_eq!(
sums,
(array![[
20.9693, 19.7502, 17.0594, 19.0700, 19.0065, 19.6254, 18.9379, 18.9275, 18.8922,
18.9505, 19.2682, 20.9411
]])
.into_dyn(),
epsilon = 1e-4
);
}
}