use std::borrow::Borrow;
use std::cmp::min;
use std::marker::PhantomData;
use crate::char_stream::InputData;
use crate::int_stream::{IntStream, IterWrapper};
use crate::token::{OwningToken, Token, TOKEN_EOF, TOKEN_INVALID_TYPE};
use crate::token_factory::TokenFactory;
use crate::token_source::TokenSource;
use better_any::{Tid, TidAble};
use std::fmt::{Debug, Formatter};
pub trait TokenStream<'input>: IntStream {
type TF: TokenFactory<'input> + 'input;
fn lt(&mut self, k: isize) -> Option<&<Self::TF as TokenFactory<'input>>::Tok>;
fn get(&self, index: isize) -> &<Self::TF as TokenFactory<'input>>::Tok;
fn get_token_source(&self) -> &dyn TokenSource<'input, TF = Self::TF>;
fn get_all_text(&self) -> String { self.get_text_from_interval(0, self.size() - 1) }
fn get_text_from_interval(&self, start: isize, stop: isize) -> String;
fn get_text_from_tokens<T: Token + ?Sized>(&self, a: &T, b: &T) -> String
where
Self: Sized,
{
self.get_text_from_interval(a.get_token_index(), b.get_token_index())
}
}
#[derive(Debug)]
pub struct TokenIter<'a, 'input: 'a, T: TokenStream<'input>>(
&'a mut T,
bool,
PhantomData<fn() -> &'input str>,
);
impl<'a, 'input: 'a, T: TokenStream<'input>> Iterator for TokenIter<'a, 'input, T> {
type Item = OwningToken;
fn next(&mut self) -> Option<Self::Item> {
if self.1 {
return None;
}
let result = self.0.lt(1).unwrap().borrow().to_owned();
if result.get_token_type() == TOKEN_EOF {
self.1 = true;
} else {
self.0.consume();
}
Some(result)
}
}
#[derive(Tid)]
pub struct UnbufferedTokenStream<'input, T: TokenSource<'input>> {
token_source: T,
pub(crate) tokens: Vec<<T::TF as TokenFactory<'input>>::Tok>,
pub(crate) current_token_index: isize,
markers_count: isize,
pub(crate) p: isize,
}
impl<'input, T: TokenSource<'input>> Debug for UnbufferedTokenStream<'input, T> {
fn fmt(&self, f: &mut Formatter<'_>) -> std::fmt::Result {
f.debug_struct("UnbufferedTokenStream")
.field("tokens", &self.tokens)
.field("current_token_index", &self.current_token_index)
.field("markers_count", &self.markers_count)
.field("p(buffer index)", &self.p)
.finish()
}
}
impl<'input, T: TokenSource<'input>> UnbufferedTokenStream<'input, T> {
pub fn iter(&mut self) -> IterWrapper<'_, Self> { IterWrapper(self) }
pub fn token_iter(&mut self) -> TokenIter<'_, 'input, Self> {
TokenIter(self, false, PhantomData)
}
pub fn new_buffered(source: T) -> Self {
let mut a = UnbufferedTokenStream::new_unbuffered(source);
a.mark();
a
}
pub fn new_unbuffered(source: T) -> Self {
UnbufferedTokenStream {
token_source: source,
tokens: vec![],
current_token_index: 0,
markers_count: 0,
p: 0,
}
}
fn sync(&mut self, want: isize) {
let need = (self.p + want - 1) - self.tokens.len() as isize + 1;
if need > 0 {
self.fill(need);
}
}
fn get_buffer_start_index(&self) -> isize { self.current_token_index - self.p }
pub(crate) fn fill(&mut self, need: isize) -> isize {
for i in 0..need {
if self.tokens.len() > 0
&& self.tokens.last().unwrap().borrow().get_token_type() == TOKEN_EOF
{
return i;
}
let token = self.token_source.next_token();
token
.borrow()
.set_token_index(self.get_buffer_start_index() + self.tokens.len() as isize);
self.tokens.push(token);
}
need
}
}
impl<'input, T: TokenSource<'input>> TokenStream<'input> for UnbufferedTokenStream<'input, T> {
type TF = T::TF;
#[inline]
fn lt(&mut self, i: isize) -> Option<&<Self::TF as TokenFactory<'input>>::Tok> {
if i == -1 {
return self.tokens.get(self.p as usize - 1);
}
self.sync(i);
self.tokens.get((self.p + i - 1) as usize)
}
#[inline]
fn get(&self, index: isize) -> &<Self::TF as TokenFactory<'input>>::Tok {
&self.tokens[(index - self.get_buffer_start_index()) as usize]
}
fn get_token_source(&self) -> &dyn TokenSource<'input, TF = Self::TF> { &self.token_source }
fn get_text_from_interval(&self, start: isize, stop: isize) -> String {
let buffer_start_index = self.get_buffer_start_index();
let buffer_stop_index = buffer_start_index + self.tokens.len() as isize - 1;
if start < buffer_start_index || stop > buffer_stop_index {
panic!(
"interval {}..={} not in token buffer window: {}..{}",
start, stop, buffer_start_index, buffer_stop_index
);
}
let a = start - buffer_start_index;
let b = stop - buffer_start_index;
let mut buf = String::new();
for i in a..(b + 1) {
let t = self.tokens[i as usize].borrow();
if t.get_token_type() == TOKEN_EOF {
break;
}
buf.extend(t.get_text().to_display().chars());
}
return buf;
}
}
impl<'input, T: TokenSource<'input>> IntStream for UnbufferedTokenStream<'input, T> {
#[inline]
fn consume(&mut self) {
if self.la(1) == TOKEN_EOF {
panic!("cannot consume EOF");
}
if self.p == self.tokens.len() as isize && self.markers_count == 0 {
self.tokens.clear();
self.p = -1;
}
self.p += 1;
self.current_token_index += 1;
self.sync(1);
}
#[inline]
fn la(&mut self, i: isize) -> isize {
self.lt(i)
.map(|t| t.borrow().get_token_type())
.unwrap_or(TOKEN_INVALID_TYPE)
}
#[inline]
fn mark(&mut self) -> isize {
self.markers_count += 1;
-self.markers_count
}
#[inline]
fn release(&mut self, marker: isize) {
assert_eq!(marker, -self.markers_count);
self.markers_count -= 1;
if self.markers_count == 0 {
if self.p > 0 {
self.tokens.drain(0..self.p as usize);
self.p = 0;
}
}
}
#[inline(always)]
fn index(&self) -> isize { self.current_token_index }
#[inline]
fn seek(&mut self, mut index: isize) {
if self.current_token_index == index {
return;
}
if index > self.current_token_index {
self.sync(index - self.current_token_index);
index = min(index, self.get_buffer_start_index() + self.size() + 1);
}
let i = index - self.get_buffer_start_index();
if i < 0 || i >= self.tokens.len() as isize {
panic!()
}
self.p = i;
self.current_token_index = index;
}
#[inline(always)]
fn size(&self) -> isize { self.tokens.len() as isize }
fn get_source_name(&self) -> String { self.token_source.get_source_name() }
}