1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
use proc_macro::TokenStream;
use proc_macro2::{Delimiter, Group, Punct, Spacing, Span, TokenStream as TokenStream2};
use quote::{quote, ToTokens, TokenStreamExt};
use scientific::Scientific;
use std::str::FromStr;
use syn::parse::{Parse, ParseStream};
use syn::{parse_macro_input, Error, LitFloat, LitInt, Token};
enum LitIntOrLitFloat {
LitInt(LitInt),
LitFloat(LitFloat),
}
impl LitIntOrLitFloat {
fn base10_digits(&self) -> &str {
match self {
LitIntOrLitFloat::LitInt(num) => num.base10_digits(),
LitIntOrLitFloat::LitFloat(num) => num.base10_digits(),
}
}
fn suffix(&self) -> &str {
match self {
LitIntOrLitFloat::LitInt(num) => num.suffix(),
LitIntOrLitFloat::LitFloat(num) => num.suffix(),
}
}
fn span(&self) -> Span {
match self {
LitIntOrLitFloat::LitInt(num) => num.span(),
LitIntOrLitFloat::LitFloat(num) => num.span(),
}
}
}
impl Parse for LitIntOrLitFloat {
fn parse(input: ParseStream) -> Result<Self, Error> {
let lookahead = input.lookahead1();
if lookahead.peek(LitInt) {
input.parse().map(LitIntOrLitFloat::LitInt)
} else if lookahead.peek(LitFloat) {
input.parse().map(LitIntOrLitFloat::LitFloat)
} else {
Err(lookahead.error())
}
}
}
struct LitScientific {
neg: Option<Token![-]>,
num: LitIntOrLitFloat,
}
impl Parse for LitScientific {
fn parse(input: ParseStream) -> Result<Self, Error> {
Ok(LitScientific {
neg: input.parse()?,
num: input.parse()?,
})
}
}
struct AsSlice<'a>(&'a [u8]);
impl<'a> ToTokens for AsSlice<'a> {
fn to_tokens(&self, stream: &mut TokenStream2) {
let mut body_stream = TokenStream2::new();
for d in self.0.iter() {
d.to_tokens(&mut body_stream);
body_stream.append(Punct::new(',', Spacing::Alone));
}
stream.append(Group::new(Delimiter::Bracket, body_stream));
}
}
#[allow(non_snake_case)]
#[proc_macro]
pub fn Scientific(item: TokenStream) -> TokenStream {
let LitScientific { neg, num } = parse_macro_input!(item as LitScientific);
if !num.suffix().is_empty() {
return TokenStream::from(Error::new(num.span(), "no suffix allowed").to_compile_error());
}
match Scientific::from_str(num.base10_digits()) {
Err(e) => TokenStream::from(Error::new(num.span(), e).to_compile_error()),
Ok(num) => {
if num.is_zero() {
quote!(scientific::Scientific::ZERO).into()
} else {
let neg = neg.is_some();
let mantissa = AsSlice(num.as_raw_mantissa());
let len = mantissa.0.len();
let exponent = num.exponent();
quote!(
{
const MANTISSA: [u8; #len] = #mantissa;
scientific::__private::unsafe_new(#neg, &MANTISSA, #exponent)
}
)
.into()
}
}
}
}