Module nlprule::tokenizer [−][src]
A tokenizer to split raw text into tokens. Tokens are assigned lemmas and part-of-speech tags by lookup from a Tagger and chunks containing information about noun / verb and grammatical case by a statistical Chunker. Tokens are disambiguated (i. e. information from the initial assignment is changed) in a rule-based way by DisambiguationRules.
Modules
chunk | A Chunker ported from OpenNLP. |
multiword | Checks if the input text contains multi-token phrases from a finite list (might contain e. g. city names) and assigns lemmas and part-of-speech tags accordingly. |
tag | A dictionary-based tagger. The raw format is tuples of the form |
Structs
Tokenizer | The complete Tokenizer doing tagging, chunking and disambiguation. |
TokenizerOptions | Options for a tokenizer. |
Functions
finalize | Finalizes the tokens by e. g. adding a specific UNKNOWN part-of-speech tag. After finalization grammatical error correction rules can be used on the tokens. |