use crate::{
input::{str::StrInput, Input},
scanner::{ScalarStyle, ScanError, Scanner, Span, Token, TokenType},
BufferedInput, Marker,
};
use std::{borrow::Cow, collections::HashMap, fmt::Display};
#[derive(Clone, Copy, PartialEq, Debug, Eq)]
enum State {
StreamStart,
ImplicitDocumentStart,
DocumentStart,
DocumentContent,
DocumentEnd,
BlockNode,
BlockSequenceFirstEntry,
BlockSequenceEntry,
IndentlessSequenceEntry,
BlockMappingFirstKey,
BlockMappingKey,
BlockMappingValue,
FlowSequenceFirstEntry,
FlowSequenceEntry,
FlowSequenceEntryMappingKey,
FlowSequenceEntryMappingValue,
FlowSequenceEntryMappingEnd(Marker),
FlowMappingFirstKey,
FlowMappingKey,
FlowMappingValue,
FlowMappingEmptyValue,
End,
}
#[derive(Clone, PartialEq, Debug, Eq)]
pub enum Event<'input> {
Nothing,
StreamStart,
StreamEnd,
DocumentStart(bool),
DocumentEnd,
Alias(
usize,
),
Scalar(
Cow<'input, str>,
ScalarStyle,
usize,
Option<Cow<'input, Tag>>,
),
SequenceStart(
usize,
Option<Cow<'input, Tag>>,
),
SequenceEnd,
MappingStart(
usize,
Option<Cow<'input, Tag>>,
),
MappingEnd,
}
#[derive(Clone, PartialEq, Debug, Eq, Ord, PartialOrd, Hash)]
pub struct Tag {
pub handle: String,
pub suffix: String,
}
impl Tag {
#[must_use]
pub fn is_yaml_core_schema(&self) -> bool {
self.handle == "tag:yaml.org,2002:"
}
}
impl Display for Tag {
fn fmt(&self, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result {
if self.handle == "!" {
write!(f, "!{}", self.suffix)
} else {
write!(f, "{}!{}", self.handle, self.suffix)
}
}
}
impl<'input> Event<'input> {
fn empty_scalar() -> Self {
Event::Scalar("~".into(), ScalarStyle::Plain, 0, None)
}
fn empty_scalar_with_anchor(anchor: usize, tag: Option<Cow<'input, Tag>>) -> Self {
Event::Scalar(Cow::default(), ScalarStyle::Plain, anchor, tag)
}
}
#[derive(Debug)]
pub struct Parser<'input, T: Input> {
scanner: Scanner<'input, T>,
states: Vec<State>,
state: State,
token: Option<Token<'input>>,
current: Option<(Event<'input>, Span)>,
anchors: HashMap<Cow<'input, str>, usize>,
anchor_id_count: usize,
tags: HashMap<String, String>,
stream_end_emitted: bool,
keep_tags: bool,
}
pub trait EventReceiver<'input> {
fn on_event(&mut self, ev: Event<'input>);
}
pub trait SpannedEventReceiver<'input> {
fn on_event(&mut self, ev: Event<'input>, span: Span);
}
impl<'input, R: EventReceiver<'input>> SpannedEventReceiver<'input> for R {
fn on_event(&mut self, ev: Event<'input>, _span: Span) {
self.on_event(ev);
}
}
pub type ParseResult<'input> = Result<(Event<'input>, Span), ScanError>;
impl<'input> Parser<'input, StrInput<'input>> {
#[must_use]
pub fn new_from_str(value: &'input str) -> Self {
debug_print!("\x1B[;31m>>>>>>>>>> New parser from str\x1B[;0m");
Parser::new(StrInput::new(value))
}
}
impl<'input, T> Parser<'input, BufferedInput<T>>
where
T: Iterator<Item = char> + 'input,
{
#[must_use]
pub fn new_from_iter(iter: T) -> Self {
debug_print!("\x1B[;31m>>>>>>>>>> New parser from iter\x1B[;0m");
Parser::new(BufferedInput::new(iter))
}
}
impl<'input, T: Input> Parser<'input, T> {
pub fn new(src: T) -> Self {
Parser {
scanner: Scanner::new(src),
states: Vec::new(),
state: State::StreamStart,
token: None,
current: None,
anchors: HashMap::new(),
anchor_id_count: 1,
tags: HashMap::new(),
stream_end_emitted: false,
keep_tags: false,
}
}
#[must_use]
pub fn keep_tags(mut self, value: bool) -> Self {
self.keep_tags = value;
self
}
pub fn peek(&mut self) -> Option<Result<&(Event<'input>, Span), ScanError>> {
if let Some(ref x) = self.current {
Some(Ok(x))
} else {
if self.stream_end_emitted {
return None;
}
match self.next_event_impl() {
Ok(token) => self.current = Some(token),
Err(e) => return Some(Err(e)),
}
self.current.as_ref().map(Ok)
}
}
pub fn next_event(&mut self) -> Option<ParseResult<'input>> {
if self.stream_end_emitted {
return None;
}
let tok = self.next_event_impl();
if matches!(tok, Ok((Event::StreamEnd, _))) {
self.stream_end_emitted = true;
}
Some(tok)
}
fn next_event_impl<'a>(&mut self) -> ParseResult<'a>
where
'input: 'a,
{
match self.current.take() {
None => self.parse(),
Some(v) => Ok(v),
}
}
fn peek_token(&mut self) -> Result<&Token, ScanError> {
match self.token {
None => {
self.token = Some(self.scan_next_token()?);
Ok(self.token.as_ref().unwrap())
}
Some(ref tok) => Ok(tok),
}
}
fn scan_next_token(&mut self) -> Result<Token<'input>, ScanError> {
let token = self.scanner.next();
match token {
None => match self.scanner.get_error() {
None => Err(ScanError::new_str(self.scanner.mark(), "unexpected eof")),
Some(e) => Err(e),
},
Some(tok) => Ok(tok),
}
}
fn fetch_token<'a>(&mut self) -> Token<'a>
where
'input: 'a,
{
self.token
.take()
.expect("fetch_token needs to be preceded by peek_token")
}
fn skip(&mut self) {
self.token = None;
}
fn pop_state(&mut self) {
self.state = self.states.pop().unwrap();
}
fn push_state(&mut self, state: State) {
self.states.push(state);
}
fn parse<'a>(&mut self) -> ParseResult<'a>
where
'input: 'a,
{
if self.state == State::End {
return Ok((Event::StreamEnd, Span::empty(self.scanner.mark())));
}
let (ev, mark) = self.state_machine()?;
Ok((ev, mark))
}
pub fn load<R: SpannedEventReceiver<'input>>(
&mut self,
recv: &mut R,
multi: bool,
) -> Result<(), ScanError> {
if !self.scanner.stream_started() {
let (ev, span) = self.next_event_impl()?;
if ev != Event::StreamStart {
return Err(ScanError::new_str(
span.start,
"did not find expected <stream-start>",
));
}
recv.on_event(ev, span);
}
if self.scanner.stream_ended() {
recv.on_event(Event::StreamEnd, Span::empty(self.scanner.mark()));
return Ok(());
}
loop {
let (ev, span) = self.next_event_impl()?;
if ev == Event::StreamEnd {
recv.on_event(ev, span);
return Ok(());
}
self.anchors.clear();
self.load_document(ev, span, recv)?;
if !multi {
break;
}
}
Ok(())
}
fn load_document<R: SpannedEventReceiver<'input>>(
&mut self,
first_ev: Event<'input>,
span: Span,
recv: &mut R,
) -> Result<(), ScanError> {
if !matches!(first_ev, Event::DocumentStart(_)) {
return Err(ScanError::new_str(
span.start,
"did not find expected <document-start>",
));
}
recv.on_event(first_ev, span);
let (ev, span) = self.next_event_impl()?;
self.load_node(ev, span, recv)?;
let (ev, mark) = self.next_event_impl()?;
assert_eq!(ev, Event::DocumentEnd);
recv.on_event(ev, mark);
Ok(())
}
fn load_node<R: SpannedEventReceiver<'input>>(
&mut self,
first_ev: Event<'input>,
span: Span,
recv: &mut R,
) -> Result<(), ScanError> {
match first_ev {
Event::Alias(..) | Event::Scalar(..) => {
recv.on_event(first_ev, span);
Ok(())
}
Event::SequenceStart(..) => {
recv.on_event(first_ev, span);
self.load_sequence(recv)
}
Event::MappingStart(..) => {
recv.on_event(first_ev, span);
self.load_mapping(recv)
}
_ => {
println!("UNREACHABLE EVENT: {first_ev:?}");
unreachable!();
}
}
}
fn load_mapping<R: SpannedEventReceiver<'input>>(
&mut self,
recv: &mut R,
) -> Result<(), ScanError> {
let (mut key_ev, mut key_mark) = self.next_event_impl()?;
while key_ev != Event::MappingEnd {
self.load_node(key_ev, key_mark, recv)?;
let (ev, mark) = self.next_event_impl()?;
self.load_node(ev, mark, recv)?;
let (ev, mark) = self.next_event_impl()?;
key_ev = ev;
key_mark = mark;
}
recv.on_event(key_ev, key_mark);
Ok(())
}
fn load_sequence<R: SpannedEventReceiver<'input>>(
&mut self,
recv: &mut R,
) -> Result<(), ScanError> {
let (mut ev, mut mark) = self.next_event_impl()?;
while ev != Event::SequenceEnd {
self.load_node(ev, mark, recv)?;
let (next_ev, next_mark) = self.next_event_impl()?;
ev = next_ev;
mark = next_mark;
}
recv.on_event(ev, mark);
Ok(())
}
fn state_machine<'a>(&mut self) -> ParseResult<'a>
where
'input: 'a,
{
debug_print!("\n\x1B[;33mParser state: {:?} \x1B[;0m", self.state);
match self.state {
State::StreamStart => self.stream_start(),
State::ImplicitDocumentStart => self.document_start(true),
State::DocumentStart => self.document_start(false),
State::DocumentContent => self.document_content(),
State::DocumentEnd => self.document_end(),
State::BlockNode => self.parse_node(true, false),
State::BlockMappingFirstKey => self.block_mapping_key(true),
State::BlockMappingKey => self.block_mapping_key(false),
State::BlockMappingValue => self.block_mapping_value(),
State::BlockSequenceFirstEntry => self.block_sequence_entry(true),
State::BlockSequenceEntry => self.block_sequence_entry(false),
State::FlowSequenceFirstEntry => self.flow_sequence_entry(true),
State::FlowSequenceEntry => self.flow_sequence_entry(false),
State::FlowMappingFirstKey => self.flow_mapping_key(true),
State::FlowMappingKey => self.flow_mapping_key(false),
State::FlowMappingValue => self.flow_mapping_value(false),
State::IndentlessSequenceEntry => self.indentless_sequence_entry(),
State::FlowSequenceEntryMappingKey => self.flow_sequence_entry_mapping_key(),
State::FlowSequenceEntryMappingValue => self.flow_sequence_entry_mapping_value(),
State::FlowSequenceEntryMappingEnd(mark) => self.flow_sequence_entry_mapping_end(mark),
State::FlowMappingEmptyValue => self.flow_mapping_value(true),
State::End => unreachable!(),
}
}
fn stream_start<'a>(&mut self) -> ParseResult<'a>
where
'input: 'a,
{
match *self.peek_token()? {
Token(span, TokenType::StreamStart(_)) => {
self.state = State::ImplicitDocumentStart;
self.skip();
Ok((Event::StreamStart, span))
}
Token(span, _) => Err(ScanError::new_str(
span.start,
"did not find expected <stream-start>",
)),
}
}
fn document_start<'a>(&mut self, implicit: bool) -> ParseResult<'a>
where
'input: 'a,
{
while let TokenType::DocumentEnd = self.peek_token()?.1 {
self.skip();
}
match *self.peek_token()? {
Token(span, TokenType::StreamEnd) => {
self.state = State::End;
self.skip();
Ok((Event::StreamEnd, span))
}
Token(
_,
TokenType::VersionDirective(..)
| TokenType::TagDirective(..)
| TokenType::DocumentStart,
) => {
self.explicit_document_start()
}
Token(span, _) if implicit => {
self.parser_process_directives()?;
self.push_state(State::DocumentEnd);
self.state = State::BlockNode;
Ok((Event::DocumentStart(false), span))
}
_ => {
self.explicit_document_start()
}
}
}
fn parser_process_directives(&mut self) -> Result<(), ScanError> {
let mut version_directive_received = false;
loop {
let mut tags = HashMap::new();
match self.peek_token()? {
Token(span, TokenType::VersionDirective(_, _)) => {
if version_directive_received {
return Err(ScanError::new_str(
span.start,
"duplicate version directive",
));
}
version_directive_received = true;
}
Token(mark, TokenType::TagDirective(handle, prefix)) => {
if tags.contains_key(&**handle) {
return Err(ScanError::new_str(mark.start, "the TAG directive must only be given at most once per handle in the same document"));
}
tags.insert(handle.to_string(), prefix.to_string());
}
_ => break,
}
self.tags = tags;
self.skip();
}
Ok(())
}
fn explicit_document_start<'a>(&mut self) -> ParseResult<'a>
where
'input: 'a,
{
self.parser_process_directives()?;
match *self.peek_token()? {
Token(mark, TokenType::DocumentStart) => {
self.push_state(State::DocumentEnd);
self.state = State::DocumentContent;
self.skip();
Ok((Event::DocumentStart(true), mark))
}
Token(span, _) => Err(ScanError::new_str(
span.start,
"did not find expected <document start>",
)),
}
}
fn document_content<'a>(&mut self) -> ParseResult<'a>
where
'input: 'a,
{
match *self.peek_token()? {
Token(
mark,
TokenType::VersionDirective(..)
| TokenType::TagDirective(..)
| TokenType::DocumentStart
| TokenType::DocumentEnd
| TokenType::StreamEnd,
) => {
self.pop_state();
Ok((Event::empty_scalar(), mark))
}
_ => self.parse_node(true, false),
}
}
fn document_end<'a>(&mut self) -> ParseResult<'a>
where
'input: 'a,
{
let mut explicit_end = false;
let span: Span = match *self.peek_token()? {
Token(span, TokenType::DocumentEnd) => {
explicit_end = true;
self.skip();
span
}
Token(span, _) => span,
};
if !self.keep_tags {
self.tags.clear();
}
if explicit_end {
self.state = State::ImplicitDocumentStart;
} else {
if let Token(span, TokenType::VersionDirective(..) | TokenType::TagDirective(..)) =
*self.peek_token()?
{
return Err(ScanError::new_str(
span.start,
"missing explicit document end marker before directive",
));
}
self.state = State::DocumentStart;
}
Ok((Event::DocumentEnd, span))
}
fn register_anchor(&mut self, name: Cow<'input, str>, _: &Span) -> usize {
let new_id = self.anchor_id_count;
self.anchor_id_count += 1;
self.anchors.insert(name, new_id);
new_id
}
fn parse_node<'a>(&mut self, block: bool, indentless_sequence: bool) -> ParseResult<'a>
where
'input: 'a,
{
let mut anchor_id = 0;
let mut tag = None;
match *self.peek_token()? {
Token(_, TokenType::Alias(_)) => {
self.pop_state();
if let Token(span, TokenType::Alias(name)) = self.fetch_token() {
match self.anchors.get(&*name) {
None => {
return Err(ScanError::new_str(
span.start,
"while parsing node, found unknown anchor",
))
}
Some(id) => return Ok((Event::Alias(*id), span)),
}
}
unreachable!()
}
Token(_, TokenType::Anchor(_)) => {
if let Token(span, TokenType::Anchor(name)) = self.fetch_token() {
anchor_id = self.register_anchor(name, &span);
if let TokenType::Tag(..) = self.peek_token()?.1 {
if let TokenType::Tag(handle, suffix) = self.fetch_token().1 {
tag = Some(self.resolve_tag(span, &handle, suffix)?);
} else {
unreachable!()
}
}
} else {
unreachable!()
}
}
Token(mark, TokenType::Tag(..)) => {
if let TokenType::Tag(handle, suffix) = self.fetch_token().1 {
tag = Some(self.resolve_tag(mark, &handle, suffix)?);
if let TokenType::Anchor(_) = &self.peek_token()?.1 {
if let Token(mark, TokenType::Anchor(name)) = self.fetch_token() {
anchor_id = self.register_anchor(name, &mark);
} else {
unreachable!()
}
}
} else {
unreachable!()
}
}
_ => {}
}
match *self.peek_token()? {
Token(mark, TokenType::BlockEntry) if indentless_sequence => {
self.state = State::IndentlessSequenceEntry;
Ok((Event::SequenceStart(anchor_id, tag), mark))
}
Token(_, TokenType::Scalar(..)) => {
self.pop_state();
if let Token(mark, TokenType::Scalar(style, v)) = self.fetch_token() {
Ok((Event::Scalar(v, style, anchor_id, tag), mark))
} else {
unreachable!()
}
}
Token(mark, TokenType::FlowSequenceStart) => {
self.state = State::FlowSequenceFirstEntry;
Ok((Event::SequenceStart(anchor_id, tag), mark))
}
Token(mark, TokenType::FlowMappingStart) => {
self.state = State::FlowMappingFirstKey;
Ok((Event::MappingStart(anchor_id, tag), mark))
}
Token(mark, TokenType::BlockSequenceStart) if block => {
self.state = State::BlockSequenceFirstEntry;
Ok((Event::SequenceStart(anchor_id, tag), mark))
}
Token(mark, TokenType::BlockMappingStart) if block => {
self.state = State::BlockMappingFirstKey;
Ok((Event::MappingStart(anchor_id, tag), mark))
}
Token(mark, _) if tag.is_some() || anchor_id > 0 => {
self.pop_state();
Ok((Event::empty_scalar_with_anchor(anchor_id, tag), mark))
}
Token(span, _) => Err(ScanError::new_str(
span.start,
"while parsing a node, did not find expected node content",
)),
}
}
fn block_mapping_key<'a>(&mut self, first: bool) -> ParseResult<'a>
where
'input: 'a,
{
if first {
let _ = self.peek_token()?;
self.skip();
}
match *self.peek_token()? {
Token(_, TokenType::Key) => {
self.skip();
if let Token(mark, TokenType::Key | TokenType::Value | TokenType::BlockEnd) =
*self.peek_token()?
{
self.state = State::BlockMappingValue;
Ok((Event::empty_scalar(), mark))
} else {
self.push_state(State::BlockMappingValue);
self.parse_node(true, true)
}
}
Token(mark, TokenType::Value) => {
self.state = State::BlockMappingValue;
Ok((Event::empty_scalar(), mark))
}
Token(mark, TokenType::BlockEnd) => {
self.pop_state();
self.skip();
Ok((Event::MappingEnd, mark))
}
Token(span, _) => Err(ScanError::new_str(
span.start,
"while parsing a block mapping, did not find expected key",
)),
}
}
fn block_mapping_value<'a>(&mut self) -> ParseResult<'a>
where
'input: 'a,
{
match *self.peek_token()? {
Token(mark, TokenType::Value) => {
self.skip();
if let Token(_, TokenType::Key | TokenType::Value | TokenType::BlockEnd) =
*self.peek_token()?
{
self.state = State::BlockMappingKey;
Ok((Event::empty_scalar(), mark))
} else {
self.push_state(State::BlockMappingKey);
self.parse_node(true, true)
}
}
Token(mark, _) => {
self.state = State::BlockMappingKey;
Ok((Event::empty_scalar(), mark))
}
}
}
fn flow_mapping_key<'a>(&mut self, first: bool) -> ParseResult<'a>
where
'input: 'a,
{
if first {
let _ = self.peek_token()?;
self.skip();
}
let span: Span = {
match *self.peek_token()? {
Token(mark, TokenType::FlowMappingEnd) => mark,
Token(mark, _) => {
if !first {
match *self.peek_token()? {
Token(_, TokenType::FlowEntry) => self.skip(),
Token(span, _) => return Err(ScanError::new_str(
span.start,
"while parsing a flow mapping, did not find expected ',' or '}'",
)),
}
}
match *self.peek_token()? {
Token(_, TokenType::Key) => {
self.skip();
if let Token(
mark,
TokenType::Value | TokenType::FlowEntry | TokenType::FlowMappingEnd,
) = *self.peek_token()?
{
self.state = State::FlowMappingValue;
return Ok((Event::empty_scalar(), mark));
}
self.push_state(State::FlowMappingValue);
return self.parse_node(false, false);
}
Token(marker, TokenType::Value) => {
self.state = State::FlowMappingValue;
return Ok((Event::empty_scalar(), marker));
}
Token(_, TokenType::FlowMappingEnd) => (),
_ => {
self.push_state(State::FlowMappingEmptyValue);
return self.parse_node(false, false);
}
}
mark
}
}
};
self.pop_state();
self.skip();
Ok((Event::MappingEnd, span))
}
fn flow_mapping_value<'a>(&mut self, empty: bool) -> ParseResult<'a>
where
'input: 'a,
{
let span: Span = {
if empty {
let Token(mark, _) = *self.peek_token()?;
self.state = State::FlowMappingKey;
return Ok((Event::empty_scalar(), mark));
}
match *self.peek_token()? {
Token(span, TokenType::Value) => {
self.skip();
match self.peek_token()?.1 {
TokenType::FlowEntry | TokenType::FlowMappingEnd => {}
_ => {
self.push_state(State::FlowMappingKey);
return self.parse_node(false, false);
}
}
span
}
Token(marker, _) => marker,
}
};
self.state = State::FlowMappingKey;
Ok((Event::empty_scalar(), span))
}
fn flow_sequence_entry<'a>(&mut self, first: bool) -> ParseResult<'a>
where
'input: 'a,
{
if first {
let _ = self.peek_token()?;
self.skip();
}
match *self.peek_token()? {
Token(mark, TokenType::FlowSequenceEnd) => {
self.pop_state();
self.skip();
return Ok((Event::SequenceEnd, mark));
}
Token(_, TokenType::FlowEntry) if !first => {
self.skip();
}
Token(span, _) if !first => {
return Err(ScanError::new_str(
span.start,
"while parsing a flow sequence, expected ',' or ']'",
));
}
_ => { }
}
match *self.peek_token()? {
Token(mark, TokenType::FlowSequenceEnd) => {
self.pop_state();
self.skip();
Ok((Event::SequenceEnd, mark))
}
Token(mark, TokenType::Key) => {
self.state = State::FlowSequenceEntryMappingKey;
self.skip();
Ok((Event::MappingStart(0, None), mark))
}
_ => {
self.push_state(State::FlowSequenceEntry);
self.parse_node(false, false)
}
}
}
fn indentless_sequence_entry<'a>(&mut self) -> ParseResult<'a>
where
'input: 'a,
{
match *self.peek_token()? {
Token(mark, TokenType::BlockEntry) => {
self.skip();
if let Token(
_,
TokenType::BlockEntry | TokenType::Key | TokenType::Value | TokenType::BlockEnd,
) = *self.peek_token()?
{
self.state = State::IndentlessSequenceEntry;
Ok((Event::empty_scalar(), mark))
} else {
self.push_state(State::IndentlessSequenceEntry);
self.parse_node(true, false)
}
}
Token(mark, _) => {
self.pop_state();
Ok((Event::SequenceEnd, mark))
}
}
}
fn block_sequence_entry<'a>(&mut self, first: bool) -> ParseResult<'a>
where
'input: 'a,
{
if first {
let _ = self.peek_token()?;
self.skip();
}
match *self.peek_token()? {
Token(mark, TokenType::BlockEnd) => {
self.pop_state();
self.skip();
Ok((Event::SequenceEnd, mark))
}
Token(mark, TokenType::BlockEntry) => {
self.skip();
if let Token(_, TokenType::BlockEntry | TokenType::BlockEnd) = *self.peek_token()? {
self.state = State::BlockSequenceEntry;
Ok((Event::empty_scalar(), mark))
} else {
self.push_state(State::BlockSequenceEntry);
self.parse_node(true, false)
}
}
Token(span, _) => Err(ScanError::new_str(
span.start,
"while parsing a block collection, did not find expected '-' indicator",
)),
}
}
fn flow_sequence_entry_mapping_key<'a>(&mut self) -> ParseResult<'a>
where
'input: 'a,
{
if let Token(mark, TokenType::Value | TokenType::FlowEntry | TokenType::FlowSequenceEnd) =
*self.peek_token()?
{
self.skip();
self.state = State::FlowSequenceEntryMappingValue;
Ok((Event::empty_scalar(), mark))
} else {
self.push_state(State::FlowSequenceEntryMappingValue);
self.parse_node(false, false)
}
}
fn flow_sequence_entry_mapping_value<'a>(&mut self) -> ParseResult<'a>
where
'input: 'a,
{
match *self.peek_token()? {
Token(_, TokenType::Value) => {
self.skip();
self.state = State::FlowSequenceEntryMappingValue;
let Token(span, ref tok) = *self.peek_token()?;
if matches!(tok, TokenType::FlowEntry | TokenType::FlowSequenceEnd) {
self.state = State::FlowSequenceEntryMappingEnd(span.end);
Ok((Event::empty_scalar(), span))
} else {
self.push_state(State::FlowSequenceEntryMappingEnd(span.end));
self.parse_node(false, false)
}
}
Token(mark, _) => {
self.state = State::FlowSequenceEntryMappingEnd(mark.end);
Ok((Event::empty_scalar(), mark))
}
}
}
#[allow(clippy::unnecessary_wraps)]
fn flow_sequence_entry_mapping_end<'a>(&mut self, mark: Marker) -> ParseResult<'a>
where
'input: 'a,
{
self.state = State::FlowSequenceEntry;
Ok((Event::MappingEnd, Span::empty(mark)))
}
fn resolve_tag(
&self,
span: Span,
handle: &str,
suffix: String,
) -> Result<Cow<'input, Tag>, ScanError> {
let tag = if handle == "!!" {
Tag {
handle: self
.tags
.get("!!")
.map_or_else(|| "tag:yaml.org,2002:".to_string(), ToString::to_string),
suffix,
}
} else if handle.is_empty() && suffix == "!" {
match self.tags.get("") {
Some(prefix) => Tag {
handle: prefix.to_string(),
suffix,
},
None => Tag {
handle: String::new(),
suffix,
},
}
} else {
let prefix = self.tags.get(handle);
if let Some(prefix) = prefix {
Tag {
handle: prefix.to_string(),
suffix,
}
} else {
if handle.len() >= 2 && handle.starts_with('!') && handle.ends_with('!') {
return Err(ScanError::new_str(span.start, "the handle wasn't declared"));
}
Tag {
handle: handle.to_string(),
suffix,
}
}
};
Ok(Cow::Owned(tag))
}
}
impl<'input, T: Input> Iterator for Parser<'input, T> {
type Item = Result<(Event<'input>, Span), ScanError>;
fn next(&mut self) -> Option<Self::Item> {
self.next_event()
}
}
#[cfg(test)]
mod test {
use super::{Event, Parser};
#[test]
fn test_peek_eq_parse() {
let s = "
a0 bb: val
a1: &x
b1: 4
b2: d
a2: 4
a3: [1, 2, 3]
a4:
- [a1, a2]
- 2
a5: *x
";
let mut p = Parser::new_from_str(s);
loop {
let event_peek = p.peek().unwrap().unwrap().clone();
let event = p.next_event().unwrap().unwrap();
assert_eq!(event, event_peek);
if event.0 == Event::StreamEnd {
break;
}
}
}
#[test]
fn test_keep_tags_across_multiple_documents() {
let text = r#"
%YAML 1.1
%TAG !t! tag:test,2024:
--- !t!1 &1
foo: "bar"
--- !t!2 &2
baz: "qux"
"#;
for x in Parser::new_from_str(text).keep_tags(true) {
let x = x.unwrap();
if let Event::MappingStart(_, tag) = x.0 {
let tag = tag.unwrap();
assert_eq!(tag.handle, "tag:test,2024:");
}
}
for x in Parser::new_from_str(text).keep_tags(false) {
if x.is_err() {
return;
}
}
panic!("Test failed, did not encounter error")
}
}