use crate::parse::keyword::Keyword;
use crate::parse::rules::*;
use crate::parse::tokenize::*;
use crate::{Error, Result};
#[derive(Clone)]
pub(crate) struct SectionRules<T: Keyword> {
rules: Vec<Option<TokenFmt<T>>>,
}
#[derive(Clone)]
enum TokVal<'a, K: Keyword> {
None,
Some([Item<'a, K>; 1]),
Multi(Vec<Item<'a, K>>),
}
impl<'a, K: Keyword> TokVal<'a, K> {
fn count(&self) -> usize {
match self {
TokVal::None => 0,
TokVal::Some(_) => 1,
TokVal::Multi(v) => v.len(),
}
}
fn first(&self) -> Option<&Item<'a, K>> {
match self {
TokVal::None => None,
TokVal::Some([t]) => Some(t),
TokVal::Multi(v) => Some(&v[0]),
}
}
fn singleton(&self) -> Option<&Item<'a, K>> {
match self {
TokVal::None => None,
TokVal::Some([t]) => Some(t),
TokVal::Multi(_) => None,
}
}
fn as_slice(&self) -> &[Item<'a, K>] {
match self {
TokVal::None => &[],
TokVal::Some(t) => &t[..],
TokVal::Multi(v) => &v[..],
}
}
fn last(&self) -> Option<&Item<'a, K>> {
match self {
TokVal::None => None,
TokVal::Some([t]) => Some(t),
TokVal::Multi(v) => Some(&v[v.len() - 1]),
}
}
}
pub struct Section<'a, T: Keyword> {
v: Vec<TokVal<'a, T>>,
first: Option<T>,
last: Option<T>,
}
impl<'a, T: Keyword> Section<'a, T> {
fn new() -> Self {
let n = T::n_vals();
let mut v = Vec::with_capacity(n);
v.resize(n, TokVal::None);
Section {
v,
first: None,
last: None,
}
}
fn tokval(&self, t: T) -> &TokVal<'a, T> {
let idx = t.idx();
&self.v[idx]
}
pub(crate) fn slice(&self, t: T) -> &[Item<'a, T>] {
self.tokval(t).as_slice()
}
pub(crate) fn get(&self, t: T) -> Option<&Item<'a, T>> {
self.tokval(t).singleton()
}
pub(crate) fn required(&self, t: T) -> Result<&Item<'a, T>> {
self.get(t).ok_or_else(|| Error::MissingToken(t.to_str()))
}
pub(crate) fn maybe<'b>(&'b self, t: T) -> MaybeItem<'b, 'a, T> {
MaybeItem::from_option(self.get(t))
}
pub(crate) fn first_item(&self) -> Option<&Item<'a, T>> {
match self.first {
None => None,
Some(t) => self.tokval(t).first(),
}
}
pub(crate) fn last_item(&self) -> Option<&Item<'a, T>> {
match self.last {
None => None,
Some(t) => self.tokval(t).last(),
}
}
fn add_tok(&mut self, t: T, item: Item<'a, T>) {
let idx = Keyword::idx(t);
if idx >= self.v.len() {
self.v.resize(idx + 1, TokVal::None);
}
let m = &mut self.v[idx];
match m {
TokVal::None => *m = TokVal::Some([item]),
TokVal::Some([x]) => {
*m = TokVal::Multi(vec![x.clone(), item]);
}
TokVal::Multi(ref mut v) => {
v.push(item);
}
};
if self.first.is_none() {
self.first = Some(t);
}
self.last = Some(t);
}
}
impl<T: Keyword> SectionRules<T> {
pub(crate) fn new() -> Self {
let n = T::n_vals();
let mut rules = Vec::with_capacity(n);
rules.resize(n, None);
SectionRules { rules }
}
pub(crate) fn add(&mut self, t: TokenFmtBuilder<T>) {
let rule: TokenFmt<_> = t.into();
let idx = rule.kwd().idx();
assert!(self.rules[idx].is_none());
self.rules[idx] = Some(rule);
}
fn parse_unverified<'a, I>(&self, tokens: &mut I, section: &mut Section<'a, T>) -> Result<()>
where
I: Iterator<Item = Result<Item<'a, T>>>,
{
for item in tokens {
let item = item?;
let tok = item.kwd();
let tok_idx = tok.idx();
if let Some(rule) = &self.rules[tok_idx] {
assert!(rule.kwd() == tok);
section.add_tok(tok, item);
rule.check_multiplicity(section.v[tok_idx].as_slice())?;
} else {
return Err(Error::UnexpectedToken(tok.to_str(), item.pos()));
}
}
Ok(())
}
fn validate<'a>(&self, s: &Section<'a, T>) -> Result<()> {
assert_eq!(s.v.len(), self.rules.len());
for (rule, t) in self.rules.iter().zip(s.v.iter()) {
match rule {
None => {
if t.count() > 0 {
unreachable!(
"This item should have been rejected earlier, in parse_unverified()"
);
}
}
Some(rule) => {
rule.check_multiplicity(t.as_slice())?;
for item in t.as_slice() {
rule.check_item(item)?;
}
}
}
}
Ok(())
}
fn validate_objects<'a>(&self, s: &Section<'a, T>, kwd: T) -> Result<()> {
for item in s.slice(kwd).iter() {
let _ = item.obj_raw()?;
}
Ok(())
}
pub(crate) fn parse<'a, I>(&self, tokens: &mut I) -> Result<Section<'a, T>>
where
I: Iterator<Item = Result<Item<'a, T>>>,
{
let mut section = Section::new();
self.parse_unverified(tokens, &mut section)?;
self.validate(§ion)?;
self.validate_objects(§ion, T::unrecognized())?;
self.validate_objects(§ion, T::ann_unrecognized())?;
Ok(section)
}
}
#[cfg(test)]
mod test {
#![allow(clippy::unwrap_used)]
use super::SectionRules;
use crate::parse::keyword::Keyword;
use crate::parse::macros::test::Fruit;
use crate::parse::tokenize::{Item, NetDocReader};
use crate::{Error, Result};
use once_cell::sync::Lazy;
static FRUIT_SALAD: Lazy<SectionRules<Fruit>> = Lazy::new(|| {
use Fruit::*;
let mut rules = SectionRules::new();
rules.add(ANN_TASTY.rule().required().args(1..=1));
rules.add(ORANGE.rule().args(1..));
rules.add(STONEFRUIT.rule().may_repeat());
rules.add(GUAVA.rule().obj_optional());
rules.add(LEMON.rule().no_args().obj_required());
rules
});
#[test]
fn parse_section() -> Result<()> {
use Fruit::*;
let s = "\
@tasty yes
orange soda
cherry cobbler
cherry pie
plum compote
guava fresh from 7 trees
-----BEGIN GUAVA MANIFESTO-----
VGhlIGd1YXZhIGVtb2ppIGlzIG5vdCBjdXJyZW50bHkgc3VwcG9ydGVkIGluI
HVuaWNvZGUgMTMuMC4gTGV0J3MgZmlnaHQgYWdhaW5zdCBhbnRpLWd1YXZhIG
JpYXMu
-----END GUAVA MANIFESTO-----
lemon
-----BEGIN LEMON-----
8J+Niw==
-----END LEMON-----
";
let mut r: NetDocReader<'_, Fruit> = NetDocReader::new(s);
let sec = FRUIT_SALAD.parse(&mut r.iter()).unwrap();
assert_eq!(sec.required(ANN_TASTY)?.arg(0), Some("yes"));
assert!(sec.get(ORANGE).is_some());
assert_eq!(sec.get(ORANGE).unwrap().args_as_str(), "soda");
let stonefruit_slice = sec.slice(STONEFRUIT);
assert_eq!(stonefruit_slice.len(), 3);
let kwds: Vec<&str> = stonefruit_slice.iter().map(Item::kwd_str).collect();
assert_eq!(kwds, &["cherry", "cherry", "plum"]);
assert_eq!(sec.maybe(GUAVA).args_as_str(), Some("fresh from 7 trees"));
assert_eq!(sec.maybe(GUAVA).parse_arg::<u32>(2).unwrap(), Some(7));
assert!(sec.maybe(GUAVA).parse_arg::<u32>(1).is_err());
assert_eq!(sec.get(GUAVA).unwrap().obj("GUAVA MANIFESTO").unwrap(),
&b"The guava emoji is not currently supported in unicode 13.0. Let's fight against anti-guava bias."[..]);
assert_eq!(
sec.get(ANN_TASTY).unwrap() as *const Item<'_, _>,
sec.first_item().unwrap() as *const Item<'_, _>
);
assert_eq!(
sec.get(LEMON).unwrap() as *const Item<'_, _>,
sec.last_item().unwrap() as *const Item<'_, _>
);
Ok(())
}
#[test]
fn rejected() {
use crate::Pos;
fn check(s: &str, e: &Error) {
let mut r: NetDocReader<'_, Fruit> = NetDocReader::new(s);
let res = FRUIT_SALAD.parse(&mut r.iter());
assert!(res.is_err());
assert_eq!(&res.err().unwrap().within(s), e);
}
check(
"orange foo\nfoobar x\n@tasty yes\n",
&Error::UnexpectedToken("<unrecognized>", Pos::from_line(2, 1)),
);
check(
"@tasty yes\norange foo\norange bar\n",
&Error::DuplicateToken("orange", Pos::from_line(3, 1)),
);
check("orange foo\n", &Error::MissingToken("@tasty"));
check(
"@tasty nope\norange\n",
&Error::TooFewArguments("orange", Pos::from_line(2, 1)),
);
check(
"@tasty yup indeed\norange normal\n",
&Error::TooManyArguments("@tasty", Pos::from_line(1, 1)),
);
check(
"@tasty yes\nlemon\norange no\n",
&Error::MissingObject("lemon", Pos::from_line(2, 1)),
);
}
}