Package opennlp.tools.tokenize
package opennlp.tools.tokenize
-
ClassDescriptionA default
TokenContextGeneratorwhich produces events for maxent decisions for tokenization.A rule based detokenizer.A basicTokenizerimplementation which performs tokenization using character classes.A thread-safe version ofTokenizerME.A cross validator fortokenizers.TheTokenizerEvaluatormeasures the performance of the givenTokenizerwith the provided referencesamples.The factory that providesTokenizerdefault implementation and resources.ATokenizerfor converting raw text into separated tokens.TheTokenizerModelis the model used by a learnableTokenizer.This class is astream filterwhich reads in string encoded samples and createssamplesout of them.This class reads thesamplesvia anIteratorand converts the samples intoeventswhich can be used by the maxent library for training.This stream formatsObjectStreamofsamplesinto whitespace separated token strings.