1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
use crate::{options::TokenizerOption, stream::CangjieTokenStream};
use jieba_rs::Jieba;
use log::trace;
use std::sync::Arc;
use tantivy::tokenizer::BoxTokenStream;
#[derive(Clone, Debug)]
pub struct CangJieTokenizer {
pub worker: Arc<Jieba>,
pub option: TokenizerOption,
}
impl Default for CangJieTokenizer {
fn default() -> Self {
CangJieTokenizer {
worker: Arc::new(Jieba::empty()),
option: TokenizerOption::Default { hmm: false },
}
}
}
impl ::tantivy::tokenizer::Tokenizer for CangJieTokenizer {
fn token_stream<'a>(&self, text: &'a str) -> BoxTokenStream<'a> {
let result = match self.option {
TokenizerOption::All => self.worker.cut_all(text),
TokenizerOption::Default { hmm: use_hmm } => self.worker.cut(text, use_hmm),
TokenizerOption::ForSearch { hmm: use_hmm } => {
self.worker.cut_for_search(text, use_hmm)
}
TokenizerOption::Unicode => {
text.chars()
.fold((0usize, vec![]), |(offset, mut result), the_char| {
result.push(&text[offset..offset + the_char.len_utf8()]);
(offset + the_char.len_utf8(), result)
})
.1
}
};
trace!("{:?}->{:?}", text, result);
BoxTokenStream::from(CangjieTokenStream::new(result))
}
}