Package opennlp.tools.tokenize
package opennlp.tools.tokenize
Contains classes related to finding token or words in a string. All
tokenizer implement the Tokenizer interface. Currently there is the
learnable
TokenizerME, the WhitespaceTokenizer and
the SimpleTokenizer which is a character class tokenizer.-
ClassDescriptionGenerate events for maxent decisions for tokenization.A Detokenizer merges tokens back to their untokenized representation.This enum contains an operation for every token to merge the tokens together to their detokenized form.The
DetokenizerEvaluatormeasures the performance of the givenDetokenizerwith the provided referenceTokenSamples.A rule based detokenizer.Performs tokenization using character classes.Interface forTokenizerMEcontext generators.The interface for tokenizers, which segment a string into its tokens.TheTokenizerEvaluatormeasures the performance of the givenTokenizerwith the provided referenceTokenSamples.The factory that providesTokenizerdefault implementations and resources.A Tokenizer for converting raw text into separated tokens.TheTokenizerModelis the model used by a learnableTokenizer.TheTokenizerStreamuses a tokenizer to tokenize the input string and outputTokenSamples.ATokenSampleis text with token spans.This class is a stream filter which reads in string encoded samples and createsTokenSamples out of them.This class reads theTokenSamples from the givenIteratorand converts theTokenSamples intoEvents which can be used by the maxent library for training.This tokenizer uses white spaces to tokenize the input text.This stream formats aTokenSamples into whitespace separated token strings.