#![allow(clippy::forget_non_drop)]
// This example exists to allow for profiling
// applications to provide details about
// the criterion benchmarks
use ress::Tokenizer;
static STRINGS: &[&str] = &[
r#""things and stuff""#,
r#"'people and places'"#,
r#""with and escaped \"""#,
r#"'another escaped \''"#,
r#""with a new \
line""#,
r#"'another new line \
hahaha'"#,
"\"sequence double quoted\\\r\nis hard\"",
"'new line sequence\\\r\nmight be harder'",
];
fn main() {
for _ in 0..1000 {
for s in STRINGS {
let d = Tokenizer::new(s).next(true).unwrap();
core::mem::forget(d);
}
}
}