use crate::error::TokenizerError;
use crate::vocab::base_vocab::{
open_protobuf_file, read_special_token_mapping_file, register_as_special_value,
swap_key_values, SpecialTokenMap,
};
use crate::vocab::Vocab;
use std::collections::{HashMap, HashSet};
use std::path::Path;
pub static FAIRSEQ_LANGUAGE_CODES: [&str; 52] = [
">>ar<<", ">>cs<<", ">>de<<", ">>en<<", ">>es<<", ">>et<<", ">>fi<<", ">>fr<<", ">>gu<<",
">>hi<<", ">>it<<", ">>ja<<", ">>kk<<", ">>ko<<", ">>lt<<", ">>lv<<", ">>my<<", ">>ne<<",
">>nl<<", ">>ro<<", ">>ru<<", ">>si<<", ">>tr<<", ">>vi<<", ">>zh<<", ">>af<<", ">>az<<",
">>bn<<", ">>fa<<", ">>he<<", ">>hr<<", ">>id<<", ">>ka<<", ">>km<<", ">>mk<<", ">>ml<<",
">>mn<<", ">>mr<<", ">>pl<<", ">>ps<<", ">>pt<<", ">>sv<<", ">>sw<<", ">>ta<<", ">>te<<",
">>th<<", ">>tl<<", ">>uk<<", ">>ur<<", ">>xh<<", ">>gl<<", ">>sl<<",
];
#[allow(clippy::upper_case_acronyms)]
#[derive(Debug, Clone)]
pub struct MBart50Vocab {
pub values: HashMap<String, i64>,
pub indices: HashMap<i64, String>,
pub special_token_map: SpecialTokenMap,
pub special_values: HashMap<String, i64>,
pub special_indices: HashMap<i64, String>,
pub language_codes_bytes: HashSet<Vec<u8>>,
}
const DEFAULT_UNK_TOKEN: &str = "<unk>";
const DEFAULT_PAD_TOKEN: &str = "<pad>";
const DEFAULT_SEP_TOKEN: &str = "</s>";
const DEFAULT_CLS_TOKEN: &str = "<s>";
const DEFAULT_EOS_TOKEN: &str = "</s>";
const DEFAULT_MASK_TOKEN: &str = "<mask>";
impl MBart50Vocab {
pub fn get_pad_value(&self) -> &str {
self.special_token_map
.cls_token
.as_deref()
.unwrap_or(DEFAULT_PAD_TOKEN)
}
pub fn get_sep_value(&self) -> &str {
self.special_token_map
.sep_token
.as_deref()
.unwrap_or(DEFAULT_SEP_TOKEN)
}
pub fn get_cls_value(&self) -> &str {
self.special_token_map
.cls_token
.as_deref()
.unwrap_or(DEFAULT_CLS_TOKEN)
}
pub fn get_eos_value(&self) -> &str {
self.special_token_map
.eos_token
.as_deref()
.unwrap_or(DEFAULT_EOS_TOKEN)
}
pub fn get_mask_value(&self) -> &str {
self.special_token_map
.mask_token
.as_deref()
.unwrap_or(DEFAULT_MASK_TOKEN)
}
}
impl Vocab for MBart50Vocab {
fn get_unknown_value(&self) -> &str {
&self.special_token_map.unk_token
}
fn values(&self) -> &HashMap<String, i64> {
&self.values
}
fn indices(&self) -> &HashMap<i64, String> {
&self.indices
}
fn special_values(&self) -> &HashMap<String, i64> {
&self.special_values
}
fn special_indices(&self) -> &HashMap<i64, String> {
&self.special_indices
}
fn from_file<P: AsRef<Path>>(path: P) -> Result<MBart50Vocab, TokenizerError> {
let mut values = HashMap::new();
let mut special_values = HashMap::new();
let special_token_map = SpecialTokenMap {
unk_token: DEFAULT_UNK_TOKEN.to_string(),
pad_token: Some(DEFAULT_PAD_TOKEN.to_string()),
bos_token: None,
sep_token: Some(DEFAULT_SEP_TOKEN.to_string()),
cls_token: Some(DEFAULT_CLS_TOKEN.to_string()),
eos_token: Some(DEFAULT_EOS_TOKEN.to_string()),
mask_token: Some(DEFAULT_MASK_TOKEN.to_string()),
additional_special_tokens: None,
};
values.insert(
special_token_map.cls_token.as_ref().unwrap().clone(),
values.len() as i64,
);
values.insert(
special_token_map.pad_token.as_ref().unwrap().clone(),
values.len() as i64,
);
values.insert(
special_token_map.eos_token.as_ref().unwrap().clone(),
values.len() as i64,
);
values.insert(special_token_map.unk_token.clone(), values.len() as i64);
let proto = open_protobuf_file(path)?;
for piece in proto.get_pieces().iter().skip(3) {
values.insert(piece.get_piece().to_owned(), values.len() as i64);
}
for language_code in FAIRSEQ_LANGUAGE_CODES.iter() {
values.insert(language_code.to_string(), values.len() as i64);
register_as_special_value(language_code, &values, &mut special_values)?;
}
values.insert(
special_token_map.mask_token.as_ref().unwrap().to_owned(),
values.len() as i64,
);
let _ = special_token_map.register_special_values(&values, &mut special_values);
let indices = swap_key_values(&values);
let special_indices = swap_key_values(&special_values);
let language_codes_bytes = FAIRSEQ_LANGUAGE_CODES
.iter()
.map(|f| f.as_bytes().to_vec())
.collect::<HashSet<Vec<u8>>>();
Ok(MBart50Vocab {
values,
indices,
special_token_map,
special_values,
special_indices,
language_codes_bytes,
})
}
fn from_file_with_special_token_mapping<P: AsRef<Path>, S: AsRef<Path>>(
path: P,
special_token_mapping_path: S,
) -> Result<Self, TokenizerError> {
let mut values = HashMap::new();
let mut special_values = HashMap::new();
let special_token_map = read_special_token_mapping_file(special_token_mapping_path)?;
if let Some(cls_token) = special_token_map.cls_token.as_ref() {
values.insert(cls_token.clone(), values.len() as i64);
}
if let Some(pad_token) = special_token_map.pad_token.as_ref() {
values.insert(pad_token.clone(), values.len() as i64);
}
if let Some(eos_token) = special_token_map.eos_token.as_ref() {
values.insert(eos_token.clone(), values.len() as i64);
}
values.insert(special_token_map.unk_token.clone(), values.len() as i64);
let proto = open_protobuf_file(path)?;
for piece in proto.get_pieces().iter().skip(3) {
values.insert(piece.get_piece().to_owned(), values.len() as i64);
}
for language_code in FAIRSEQ_LANGUAGE_CODES.iter() {
values.insert(language_code.to_string(), values.len() as i64);
register_as_special_value(language_code, &values, &mut special_values)?;
}
values.insert(
special_token_map.mask_token.as_ref().unwrap().to_owned(),
values.len() as i64,
);
let _ = special_token_map.register_special_values(&values, &mut special_values);
let indices = swap_key_values(&values);
let special_indices = swap_key_values(&special_values);
let language_codes_bytes = FAIRSEQ_LANGUAGE_CODES
.iter()
.map(|f| f.as_bytes().to_vec())
.collect::<HashSet<Vec<u8>>>();
Ok(MBart50Vocab {
values,
indices,
special_token_map,
special_values,
special_indices,
language_codes_bytes,
})
}
fn from_values_and_special_token_map(
values: HashMap<String, i64>,
special_token_map: SpecialTokenMap,
) -> Result<Self, TokenizerError>
where
Self: Sized,
{
let mut special_values = HashMap::new();
special_token_map.register_special_values(&values, &mut special_values)?;
let indices = swap_key_values(&values);
let special_indices = swap_key_values(&special_values);
let language_codes_bytes = FAIRSEQ_LANGUAGE_CODES
.iter()
.map(|f| f.as_bytes().to_vec())
.collect::<HashSet<Vec<u8>>>();
Ok(Self {
values,
indices,
special_token_map,
special_values,
special_indices,
language_codes_bytes,
})
}
fn token_to_id(&self, token: &str) -> i64 {
self._token_to_id(
token,
&self.values,
&self.special_values,
self.get_unknown_value(),
)
}
fn id_to_token(&self, id: &i64) -> String {
self._id_to_token(
id,
&self.indices,
&self.special_indices,
self.get_unknown_value(),
)
}
}