1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
use proc_macro::{TokenStream};
use syn::parse_macro_input;
use proc_macro_error::{proc_macro_error, abort};
use quote::quote;
mod lexer;
mod parser;
mod common;
mod discriminants;
use common::*;
#[proc_macro_error]
#[proc_macro_attribute]
pub fn lexer(args: TokenStream, input: TokenStream) -> TokenStream {
let args = parse_macro_input!(args as syn::Meta);
let lexer_ident = match get_lexer_ident(&args) {
Some(id) => id,
None => abort!(args, "lexer name must be specified")
};
let input = parse_macro_input!(input as syn::ItemEnum);
match lexer::lexer(lexer_ident, input) {
Ok(s) => s.into(),
Err(ParceMacroError(who, message)) => abort!(who, message)
}
}
#[proc_macro_error]
#[proc_macro_attribute]
pub fn parser(args: TokenStream, input: TokenStream) -> TokenStream {
let args = parse_macro_input!(args as syn::Meta);
let lexer_ident = match get_lexer_path(&args) {
Some(id) => id,
None => abort!(args, "lexer name must be specified")
};
let input = parse_macro_input!(input as syn::ItemEnum);
match parser::parser(lexer_ident, input) {
Ok(s) => s.into(),
Err(ParceMacroError(who, message)) => abort!(who, message)
}
}
#[proc_macro_derive(RemoveLexerAttributes, attributes(skip, frag, set_mode, mode, modes))]
pub fn lex_attributes(_input: TokenStream) -> TokenStream {
(quote! {}).into()
}