Version 1.6.0
Copyright © 2011, 2024 The Apache Software Foundation
License and Disclaimer. The ASF licenses this documentation to you under the Apache License, Version 2.0 (the "License"); you may not use this documentation except in compliance with the License. You may obtain a copy of the License at
Unless required by applicable law or agreed to in writing, this documentation and its contents are distributed under the License on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License.
Table of Contents
List of Tables
Table of Contents
The Apache OpenNLP library is a machine learning based toolkit for the processing of natural language text. It supports the most common NLP tasks, such as tokenization, sentence segmentation, part-of-speech tagging, named entity extraction, chunking, parsing, and coreference resolution. These tasks are usually required to build more advanced text processing services. OpenNLP also included maximum entropy and perceptron based machine learning.
The goal of the OpenNLP project will be to create a mature toolkit for the abovementioned tasks. An additional goal is to provide a large number of pre-built models for a variety of languages, as well as the annotated text resources that those models are derived from.
The Apache OpenNLP library contains several components, enabling one to build a full natural language processing pipeline. These components include: sentence detector, tokenizer, name finder, document categorizer, part-of-speech tagger, chunker, parser, coreference resolution. Components contain parts which enable one to execute the respective natural language processing task, to train a model and often also to evaluate a model. Each of these facilities is accessible via its application program interface (API). In addition, a command line interface (CLI) is provided for convenience of experiments and training.
OpenNLP components have similar APIs. Normally, to execute a task, one should provide a model and an input.
A model is usually loaded by providing a FileInputStream with a model to a constructor of the model class:
InputStream modelIn = new FileInputStream("lang-model-name.bin"); try { SomeModel model = new SomeModel(modelIn); } catch (IOException e) { //handle the exception } finally { if (null != modelIn) { try { modelIn.close(); } catch (IOException e) { } } }
After the model is loaded the tool itself can be instantiated.
ToolName toolName = new ToolName(model);
After the tool is instantiated, the processing task can be executed. The input and the output formats are specific to the tool, but often the output is an array of String, and the input is a String or an array of String.
String output[] = toolName.executeTask("This is a sample text.");
OpenNLP provides a command line script, serving as a unique entry point to all included tools. The script is located in the bin directory of OpenNLP binary distribution. Included are versions for Windows: opennlp.bat and Linux or compatible systems: opennlp.
OpenNLP script uses JAVA_CMD and JAVA_HOME variables to determine which command to use to execute Java virtual machine.
OpenNLP script uses OPENNLP_HOME variable to determine the location of the binary distribution of OpenNLP. It is recommended to point this variable to the binary distribution of current OpenNLP version and update PATH variable to include $OPENNLP_HOME/bin or %OPENNLP_HOME%\bin.
Such configuration allows calling OpenNLP conveniently. Examples below suppose this configuration has been done.
Apache OpenNLP provides a common command line script to access all its tools:
$ opennlp
This script prints current version of the library and lists all available tools:
OpenNLP <VERSION>. Usage: opennlp TOOL where TOOL is one of: Doccat learnable document categorizer DoccatTrainer trainer for the learnable document categorizer DoccatConverter converts leipzig data format to native OpenNLP format DictionaryBuilder builds a new dictionary SimpleTokenizer character class tokenizer TokenizerME learnable tokenizer TokenizerTrainer trainer for the learnable tokenizer TokenizerMEEvaluator evaluator for the learnable tokenizer TokenizerCrossValidator K-fold cross validator for the learnable tokenizer TokenizerConverter converts foreign data formats (namefinder,conllx,pos) to native OpenNLP format DictionaryDetokenizer SentenceDetector learnable sentence detector SentenceDetectorTrainer trainer for the learnable sentence detector SentenceDetectorEvaluator evaluator for the learnable sentence detector SentenceDetectorCrossValidator K-fold cross validator for the learnable sentence detector SentenceDetectorConverter converts foreign data formats (namefinder,conllx,pos) to native OpenNLP format TokenNameFinder learnable name finder TokenNameFinderTrainer trainer for the learnable name finder TokenNameFinderEvaluator Measures the performance of the NameFinder model with the reference data TokenNameFinderCrossValidator K-fold cross validator for the learnable Name Finder TokenNameFinderConverter converts foreign data formats (bionlp2004,conll03,conll02,ad) to native OpenNLP format CensusDictionaryCreator Converts 1990 US Census names into a dictionary POSTagger learnable part of speech tagger POSTaggerTrainer trains a model for the part-of-speech tagger POSTaggerEvaluator Measures the performance of the POS tagger model with the reference data POSTaggerCrossValidator K-fold cross validator for the learnable POS tagger POSTaggerConverter converts conllx data format to native OpenNLP format ChunkerME learnable chunker ChunkerTrainerME trainer for the learnable chunker ChunkerEvaluator Measures the performance of the Chunker model with the reference data ChunkerCrossValidator K-fold cross validator for the chunker ChunkerConverter converts ad data format to native OpenNLP format Parser performs full syntactic parsing ParserTrainer trains the learnable parser ParserEvaluator Measures the performance of the Parser model with the reference data BuildModelUpdater trains and updates the build model in a parser model CheckModelUpdater trains and updates the check model in a parser model TaggerModelReplacer replaces the tagger model in a parser model All tools print help when invoked with help parameter Example: opennlp SimpleTokenizer help
OpenNLP tools have similar command line structure and options. To discover tool options, run it with no parameters:
$ opennlp ToolName
The tool will output two blocks of help.
The first block describes the general structure of this tool command line:
Usage: opennlp TokenizerTrainer[.namefinder|.conllx|.pos] [-abbDict path] ... -model modelFile ...
The general structure of this tool command line includes the obligatory tool name (TokenizerTrainer), the optional format parameters ([.namefinder|.conllx|.pos]), the optional parameters ([-abbDict path] ...), and the obligatory parameters (-model modelFile ...).
The format parameters enable direct processing of non-native data without conversion. Each format might have its own parameters, which are displayed if the tool is executed without or with help parameter:
$ opennlp TokenizerTrainer.conllx help
Usage: opennlp TokenizerTrainer.conllx [-abbDict path] [-alphaNumOpt isAlphaNumOpt] ... Arguments description: -abbDict path abbreviation dictionary in XML format. ...
To switch the tool to a specific format, add a dot and the format name after the tool name:
$ opennlp TokenizerTrainer.conllx -model en-pos.bin ...
The second block of the help message describes the individual arguments:
Arguments description: -type maxent|perceptron|perceptron_sequence The type of the token name finder model. One of maxent|perceptron|perceptron_sequence. -dict dictionaryPath The XML tag dictionary file ...
Most tools for processing need to be provided at least a model:
$ opennlp ToolName lang-model-name.bin
When tool is executed this way, the model is loaded and the tool is waiting for the input from standard input. This input is processed and printed to standard output.
Alternative, or one should say, most commonly used way is to use console input and output redirection options to provide also an input and an output files:
$ opennlp ToolName lang-model-name.bin < input.txt > output.txt
Most tools for model training need to be provided first a model name, optionally some training options (such as model type, number of iterations), and then the data.
A model name is just a file name.
Training options often include number of iterations, cutoff, abbreviations dictionary or something else. Sometimes it is possible to provide these options via training options file. In this case these options are ignored and the ones from the file are used.
For the data one has to specify the location of the data (filename) and often language and encoding.
A generic example of a command line to launch a tool trainer might be:
$ opennlp ToolNameTrainer -model en-model-name.bin -lang en -data input.train -encoding UTF-8
or with a format:
$ opennlp ToolNameTrainer.conll03 -model en-model-name.bin -lang en -data input.train \ -types per -encoding UTF-8
Most tools for model evaluation are similar to those for task execution, and need to be provided fist a model name, optionally some evaluation options (such as whether to print misclassified samples), and then the test data. A generic example of a command line to launch an evaluation tool might be:
$ opennlp ToolNameEvaluator -model en-model-name.bin -lang en -data input.test -encoding UTF-8
Table of Contents
The OpenNLP Sentence Detector can detect that a punctuation character marks the end of a sentence or not. In this sense a sentence is defined as the longest white space trimmed character sequence between two punctuation marks. The first and last sentence make an exception to this rule. The first non whitespace character is assumed to be the begin of a sentence, and the last non whitespace character is assumed to be a sentence end. The sample text below should be segmented into its sentences.
Pierre Vinken, 61 years old, will join the board as a nonexecutive director Nov. 29. Mr. Vinken is chairman of Elsevier N.V., the Dutch publishing group. Rudolph Agnew, 55 years old and former chairman of Consolidated Gold Fields PLC, was named a director of this British industrial conglomerate.
After detecting the sentence boundaries each sentence is written in its own line.
Pierre Vinken, 61 years old, will join the board as a nonexecutive director Nov. 29. Mr. Vinken is chairman of Elsevier N.V., the Dutch publishing group. Rudolph Agnew, 55 years old and former chairman of Consolidated Gold Fields PLC, was named a director of this British industrial conglomerate.
Usually Sentence Detection is done before the text is tokenized and that's the way the pre-trained models on the web site are trained, but it is also possible to perform tokenization first and let the Sentence Detector process the already tokenized text. The OpenNLP Sentence Detector cannot identify sentence boundaries based on the contents of the sentence. A prominent example is the first sentence in an article where the title is mistakenly identified to be the first part of the first sentence. Most components in OpenNLP expect input which is segmented into sentences.
The easiest way to try out the Sentence Detector is the command line tool. The tool is only intended for demonstration and testing. Download the english sentence detector model and start the Sentence Detector Tool with this command:
$ opennlp SentenceDetector en-sent.bin
Just copy the sample text from above to the console. The Sentence Detector will read it and echo one sentence per line to the console. Usually the input is read from a file and the output is redirected to another file. This can be achieved with the following command.
$ opennlp SentenceDetector en-sent.bin < input.txt > output.txt
For the english sentence model from the website the input text should not be tokenized.
The Sentence Detector can be easily integrated into an application via its API. To instantiate the Sentence Detector the sentence model must be loaded first.
InputStream modelIn = new FileInputStream("en-sent.bin"); try { SentenceModel model = new SentenceModel(modelIn); } catch (IOException e) { e.printStackTrace(); } finally { if (modelIn != null) { try { modelIn.close(); } catch (IOException e) { } } }
After the model is loaded the SentenceDetectorME can be instantiated.
SentenceDetectorME sentenceDetector = new SentenceDetectorME(model);
The Sentence Detector can output an array of Strings, where each String is one sentence.
String sentences[] = sentenceDetector.sentDetect(" First sentence. Second sentence. ");
The result array now contains two entries. The first String is "First sentence." and the second String is "Second sentence." The whitespace before, between and after the input String is removed. The API also offers a method which simply returns the span of the sentence in the input string.
Span sentences[] = sentenceDetector.sentPosDetect(" First sentence. Second sentence. ");
The result array again contains two entries. The first span beings at index 2 and ends at 17. The second span begins at 18 and ends at 34. The utility method Span.getCoveredText can be used to create a substring which only covers the chars in the span.
OpenNLP has a command line tool which is used to train the models available from the model download page on various corpora. The data must be converted to the OpenNLP Sentence Detector training format. Which is one sentence per line. An empty line indicates a document boundary. In case the document boundary is unknown, its recommended to have an empty line every few ten sentences. Exactly like the output in the sample above. Usage of the tool:
$ opennlp SentenceDetectorTrainer Usage: opennlp SentenceDetectorTrainer[.namefinder|.conllx|.pos] [-abbDict path] \ [-params paramsFile] [-iterations num] [-cutoff num] -model modelFile \ -lang language -data sampleData [-encoding charsetName] Arguments description: -abbDict path abbreviation dictionary in XML format. -params paramsFile training parameters file. -iterations num number of training iterations, ignored if -params is used. -cutoff num minimal number of times a feature must be seen, ignored if -params is used. -model modelFile output model file. -lang language language which is being processed. -data sampleData data to be used, usually a file name. -encoding charsetName encoding for reading and writing text, if absent the system default is used.
To train an English sentence detector use the following command:
$ opennlp SentenceDetectorTrainer -model en-sent.bin -lang en -data en-sent.train -encoding UTF-8
It should produce the following output:
Indexing events using cutoff of 5 Computing event counts... done. 4883 events Indexing... done. Sorting and merging events... done. Reduced 4883 events to 2945. Done indexing. Incorporating indexed data for training... done. Number of Event Tokens: 2945 Number of Outcomes: 2 Number of Predicates: 467 ...done. Computing model parameters... Performing 100 iterations. 1: .. loglikelihood=-3384.6376826743144 0.38951464263772273 2: .. loglikelihood=-2191.9266688597672 0.9397911120212984 3: .. loglikelihood=-1645.8640771555981 0.9643661683391358 4: .. loglikelihood=-1340.386303774519 0.9739913987302887 5: .. loglikelihood=-1148.4141548519624 0.9748105672742167 ...<skipping a bunch of iterations>... 95: .. loglikelihood=-288.25556805874436 0.9834118369854598 96: .. loglikelihood=-287.2283680343481 0.9834118369854598 97: .. loglikelihood=-286.2174830344526 0.9834118369854598 98: .. loglikelihood=-285.222486981048 0.9834118369854598 99: .. loglikelihood=-284.24296917223916 0.9834118369854598 100: .. loglikelihood=-283.2785335773966 0.9834118369854598 Wrote sentence detector model. Path: en-sent.bin
The Sentence Detector also offers an API to train a new sentence detection model. Basically three steps are necessary to train it:
The application must open a sample data stream
Call the SentenceDetectorME.train method
Save the SentenceModel to a file or directly use it
The following sample code illustrates these steps:
Charset charset = Charset.forName("UTF-8"); ObjectStream<String> lineStream = new PlainTextByLineStream(new FileInputStream("en-sent.train"), charset); ObjectStream<SentenceSample> sampleStream = new SentenceSampleStream(lineStream); SentenceModel model; try { model = SentenceDetectorME.train("en", sampleStream, true, null, TrainingParameters.defaultParams()); } finally { sampleStream.close(); } OutputStream modelOut = null; try { modelOut = new BufferedOutputStream(new FileOutputStream(modelFile)); model.serialize(modelOut); } finally { if (modelOut != null) modelOut.close(); }
The command shows how the evaluator tool can be run:
$ opennlp SentenceDetectorEvaluator -model en-sent.bin -lang en -data en-sent.eval -encoding UTF-8 Loading model ... done Evaluating ... done Precision: 0.9465737514518002 Recall: 0.9095982142857143 F-Measure: 0.9277177006260672
The en-sent.eval file has the same format as the training data.
Table of Contents
The OpenNLP Tokenizers segment an input character sequence into tokens. Tokens are usually words, punctuation, numbers, etc.
Pierre Vinken, 61 years old, will join the board as a nonexecutive director Nov. 29. Mr. Vinken is chairman of Elsevier N.V., the Dutch publishing group. Rudolph Agnew, 55 years old and former chairman of Consolidated Gold Fields PLC, was named a director of this British industrial conglomerate.
The following result shows the individual tokens in a whitespace separated representation.
Pierre Vinken , 61 years old , will join the board as a nonexecutive director Nov. 29 . Mr. Vinken is chairman of Elsevier N.V. , the Dutch publishing group . Rudolph Agnew , 55 years old and former chairman of Consolidated Gold Fields PLC , was named a nonexecutive director of this British industrial conglomerate . A form of asbestos once used to make Kent cigarette filters has caused a high percentage of cancer deaths among a group of workers exposed to it more than 30 years ago , researchers reported .
OpenNLP offers multiple tokenizer implementations:
Whitespace Tokenizer - A whitespace tokenizer, non whitespace sequences are identified as tokens
Simple Tokenizer - A character class tokenizer, sequences of the same character class are tokens
Learnable Tokenizer - A maximum entropy tokenizer, detects token boundaries based on probability model
Most part-of-speech taggers, parsers and so on, work with text tokenized in this manner. It is important to ensure that your tokenizer produces tokens of the type expected by your later text processing components.
With OpenNLP (as with many systems), tokenization is a two-stage process: first, sentence boundaries are identified, then tokens within each sentence are identified.
The easiest way to try out the tokenizers are the command line tools. The tools are only intended for demonstration and testing.
There are two tools, one for the Simple Tokenizer and one for the learnable tokenizer. A command line tool the for the Whitespace Tokenizer does not exist, because the whitespace separated output would be identical to the input.
The following command shows how to use the Simple Tokenizer Tool.
$ opennlp SimpleTokenizer
To use the learnable tokenizer download the english token model from our website.
$ opennlp TokenizerME en-token.bin
To test the tokenizer copy the sample from above to the console. The whitespace separated tokens will be written back to the console.
Usually the input is read from a file and written to a file.
$ opennlp TokenizerME en-token.bin < article.txt > article-tokenized.txt
It can be done in the same way for the Simple Tokenizer.
Since most text comes truly raw and doesn't have sentence boundaries and such, its possible to create a pipe which first performs sentence boundary detection and tokenization. The following sample illustrates that.
$ opennlp SentenceDetector sentdetect.model < article.txt | opennlp TokenizerME tokenize.model | more Loading model ... Loading model ... done done Showa Shell gained 20 to 1,570 and Mitsubishi Oil rose 50 to 1,500. Sumitomo Metal Mining fell five yen to 692 and Nippon Mining added 15 to 960 . Among other winners Wednesday was Nippon Shokubai , which was up 80 at 2,410 . Marubeni advanced 11 to 890 . London share prices were bolstered largely by continued gains on Wall Street and technical factors affecting demand for London 's blue-chip stocks . ...etc...
Of course this is all on the command line. Many people use the models directly in their Java code by creating SentenceDetector and Tokenizer objects and calling their methods as appropriate. The following section will explain how the Tokenizers can be used directly from java.
The Tokenizers can be integrated into an application by the defined API. The shared instance of the WhitespaceTokenizer can be retrieved from a static field WhitespaceTokenizer.INSTANCE. The shared instance of the SimpleTokenizer can be retrieved in the same way from SimpleTokenizer.INSTANCE. To instantiate the TokenizerME (the learnable tokenizer) a Token Model must be created first. The following code sample shows how a model can be loaded.
InputStream modelIn = new FileInputStream("en-token.bin"); try { TokenizerModel model = new TokenizerModel(modelIn); } catch (IOException e) { e.printStackTrace(); } finally { if (modelIn != null) { try { modelIn.close(); } catch (IOException e) { } } }
After the model is loaded the TokenizerME can be instantiated.
Tokenizer tokenizer = new TokenizerME(model);
The tokenizer offers two tokenize methods, both expect an input String object which contains the untokenized text. If possible it should be a sentence, but depending on the training of the learnable tokenizer this is not required. The first returns an array of Strings, where each String is one token.
String tokens[] = tokenizer.tokenize("An input sample sentence.");
The output will be an array with these tokens.
"An", "input", "sample", "sentence", "."
The second method, tokenizePos returns an array of Spans, each Span contain the begin and end character offsets of the token in the input String.
Span tokenSpans[] = tokenizer.tokenizePos("An input sample sentence.");
The tokenSpans array now contain 5 elements. To get the text for one span call Span.getCoveredText which takes a span and the input text. The TokenizerME is able to output the probabilities for the detected tokens. The getTokenProbabilities method must be called directly after one of the tokenize methods was called.
TokenizerME tokenizer = ... String tokens[] = tokenizer.tokenize(...); double tokenProbs[] = tokenizer.getTokenProbabilities();
The tokenProbs array now contains one double value per token, the value is between 0 and 1, where 1 is the highest possible probability and 0 the lowest possible probability.
OpenNLP has a command line tool which is used to train the models available from the model download page on various corpora. The data can be converted to the OpenNLP Tokenizer training format or used directly. The OpenNLP format contains one sentence per line. Tokens are either separated by a whitespace or by a special <SPLIT> tag. The following sample shows the sample from above in the correct format.
Pierre Vinken<SPLIT>, 61 years old<SPLIT>, will join the board as a nonexecutive director Nov. 29<SPLIT>. Mr. Vinken is chairman of Elsevier N.V.<SPLIT>, the Dutch publishing group<SPLIT>. Rudolph Agnew<SPLIT>, 55 years old and former chairman of Consolidated Gold Fields PLC<SPLIT>, was named a nonexecutive director of this British industrial conglomerate<SPLIT>.
Usage of the tool:
$ opennlp TokenizerTrainer Usage: opennlp TokenizerTrainer[.namefinder|.conllx|.pos] [-abbDict path] \ [-alphaNumOpt isAlphaNumOpt] [-params paramsFile] [-iterations num] \ [-cutoff num] -model modelFile -lang language -data sampleData \ [-encoding charsetName] Arguments description: -abbDict path abbreviation dictionary in XML format. -alphaNumOpt isAlphaNumOpt Optimization flag to skip alpha numeric tokens for further tokenization -params paramsFile training parameters file. -iterations num number of training iterations, ignored if -params is used. -cutoff num minimal number of times a feature must be seen, ignored if -params is used. -model modelFile output model file. -lang language language which is being processed. -data sampleData data to be used, usually a file name. -encoding charsetName encoding for reading and writing text, if absent the system default is used.
To train the english tokenizer use the following command:
$ opennlp TokenizerTrainer -model en-token.bin -alphaNumOpt -lang en -data en-token.train -encoding UTF-8 Indexing events using cutoff of 5 Computing event counts... done. 262271 events Indexing... done. Sorting and merging events... done. Reduced 262271 events to 59060. Done indexing. Incorporating indexed data for training... done. Number of Event Tokens: 59060 Number of Outcomes: 2 Number of Predicates: 15695 ...done. Computing model parameters... Performing 100 iterations. 1: .. loglikelihood=-181792.40419263614 0.9614292087192255 2: .. loglikelihood=-34208.094253153664 0.9629238459456059 3: .. loglikelihood=-18784.123872910015 0.9729211388220581 4: .. loglikelihood=-13246.88162585859 0.9856103038460219 5: .. loglikelihood=-10209.262670265718 0.9894422181636552 ...<skipping a bunch of iterations>... 95: .. loglikelihood=-769.2107474529454 0.999511955191386 96: .. loglikelihood=-763.8891914534009 0.999511955191386 97: .. loglikelihood=-758.6685383254891 0.9995157680414533 98: .. loglikelihood=-753.5458314695236 0.9995157680414533 99: .. loglikelihood=-748.5182305519613 0.9995157680414533 100: .. loglikelihood=-743.5830058068038 0.9995157680414533 Wrote tokenizer model. Path: en-token.bin
The Tokenizer offers an API to train a new tokenization model. Basically three steps are necessary to train it:
The application must open a sample data stream
Call the TokenizerME.train method
Save the TokenizerModel to a file or directly use it
The following sample code illustrates these steps:
Charset charset = Charset.forName("UTF-8"); ObjectStream<String> lineStream = new PlainTextByLineStream(new FileInputStream("en-sent.train"), charset); ObjectStream<TokenSample> sampleStream = new TokenSampleStream(lineStream); TokenizerModel model; try { model = TokenizerME.train("en", sampleStream, true, TrainingParameters.defaultParams()); } finally { sampleStream.close(); } OutputStream modelOut = null; try { modelOut = new BufferedOutputStream(new FileOutputStream(modelFile)); model.serialize(modelOut); } finally { if (modelOut != null) modelOut.close(); }
Detokenizing is simple the opposite of tokenization, the original non-tokenized string should be constructed out of a token sequence. The OpenNLP implementation was created to undo the tokenization of training data for the tokenizer. It can also be used to undo the tokenization of such a trained tokenizer. The implementation is strictly rule based and defines how tokens should be attached to a sentence wise character sequence.
The rule dictionary assign to every token an operation which describes how it should be attached to one continuous character sequence.
The following rules can be assigned to a token:
MERGE_TO_LEFT - Merges the token to the left side.
MERGE_TO_RIGHT - Merges the token to the right side.
RIGHT_LEFT_MATCHING - Merges the token to the right side on first occurrence and to the left side on second occurrence.
The following sample will illustrate how the detokenizer with a small rule dictionary (illustration format, not the xml data format):
. MERGE_TO_LEFT " RIGHT_LEFT_MATCHING
The dictionary should be used to de-tokenize the following whitespace tokenized sentence:
He said " This is a test " .
The tokens would get these tags based on the dictionary:
He -> NO_OPERATION said -> NO_OPERATION " -> MERGE_TO_RIGHT This -> NO_OPERATION is -> NO_OPERATION a -> NO_OPERATION test -> NO_OPERATION " -> MERGE_TO_LEFT . -> MERGE_TO_LEFT
That will result in the following character sequence:
He said "This is a test".
TODO: Add documentation about the dictionary format and how to use the API. Contributions are welcome.
TODO: Write documentation about the detokenizer api. Any contributions are very welcome. If you want to contribute please contact us on the mailing list or comment on the jira issue OPENNLP-216.
TODO: Write documentation about the detokenizer dictionary. Any contributions are very welcome. If you want to contribute please contact us on the mailing list or comment on the jira issue OPENNLP-217.
Table of Contents
The Name Finder can detect named entities and numbers in text. To be able to detect entities the Name Finder needs a model. The model is dependent on the language and entity type it was trained for. The OpenNLP projects offers a number of pre-trained name finder models which are trained on various freely available corpora. They can be downloaded at our model download page. To find names in raw text the text must be segmented into tokens and sentences. A detailed description is given in the sentence detector and tokenizer tutorial. It is important that the tokenization for the training data and the input text is identical.
The easiest way to try out the Name Finder is the command line tool. The tool is only intended for demonstration and testing. Download the English person model and start the Name Finder Tool with this command:
$ opennlp TokenNameFinder en-ner-person.bin
The name finder now reads a tokenized sentence per line from stdin, an empty line indicates a document boundary and resets the adaptive feature generators. Just copy this text to the terminal:
Pierre Vinken , 61 years old , will join the board as a nonexecutive director Nov. 29 . Mr . Vinken is chairman of Elsevier N.V. , the Dutch publishing group . Rudolph Agnew , 55 years old and former chairman of Consolidated Gold Fields PLC , was named a director of this British industrial conglomerate .
the name finder will now output the text with markup for person names:
<START:person> Pierre Vinken <END> , 61 years old , will join the board as a nonexecutive director Nov. 29 . Mr . <START:person> Vinken <END> is chairman of Elsevier N.V. , the Dutch publishing group . <START:person> Rudolph Agnew <END> , 55 years old and former chairman of Consolidated Gold Fields PLC , was named a director of this British industrial conglomerate .
To use the Name Finder in a production system it is strongly recommended to embed it directly into the application instead of using the command line interface. First the name finder model must be loaded into memory from disk or an other source. In the sample below it is loaded from disk.
InputStream modelIn = new FileInputStream("en-ner-person.bin"); try { TokenNameFinderModel model = new TokenNameFinderModel(modelIn); } catch (IOException e) { e.printStackTrace(); } finally { if (modelIn != null) { try { modelIn.close(); } catch (IOException e) { } } }
There is a number of reasons why the model loading can fail:
Issues with the underlying I/O
The version of the model is not compatible with the OpenNLP version
The model is loaded into the wrong component, for example a tokenizer model is loaded with TokenNameFinderModel class.
The model content is not valid for some other reason
After the model is loaded the NameFinderME can be instantiated.
NameFinderME nameFinder = new NameFinderME(model);
The initialization is now finished and the Name Finder can be used. The NameFinderME class is not thread safe, it must only be called from one thread. To use multiple threads multiple NameFinderME instances sharing the same model instance can be created. The input text should be segmented into documents, sentences and tokens. To perform entity detection an application calls the find method for every sentence in the document. After every document clearAdaptiveData must be called to clear the adaptive data in the feature generators. Not calling clearAdaptiveData can lead to a sharp drop in the detection rate after a few documents. The following code illustrates that:
for (String document[][] : documents) { for (String[] sentence : document) { Span nameSpans[] = nameFinder.find(sentence); // do something with the names } nameFinder.clearAdaptiveData() }
the following snippet shows a call to find
String sentence[] = new String[]{ "Pierre", "Vinken", "is", "61", "years" "old", "." }; Span nameSpans[] = nameFinder.find(sentence);
The nameSpans arrays contains now exactly one Span which marks the name Pierre Vinken. The elements between the begin and end offsets are the name tokens. In this case the begin offset is 0 and the end offset is 2. The Span object also knows the type of the entity. In this case it is person (defined by the model). It can be retrieved with a call to Span.getType(). Additionally to the statistical Name Finder, OpenNLP also offers a dictionary and a regular expression name finder implementation.
TODO: Explain how to retrieve probs from the name finder for names and for non recognized names
The pre-trained models might not be available for a desired language, can not detect important entities or the performance is not good enough outside the news domain. These are the typical reason to do custom training of the name finder on a new corpus or on a corpus which is extended by private training data taken from the data which should be analyzed.
OpenNLP has a command line tool which is used to train the models available from the model download page on various corpora.
The data can be converted to the OpenNLP name finder training format. Which is one sentence per line. Some other formats are available as well. The sentence must be tokenized and contain spans which mark the entities. Documents are separated by empty lines which trigger the reset of the adaptive feature generators. A training file can contain multiple types. If the training file contains multiple types the created model will also be able to detect these multiple types. For now it is recommended to only train single type models, since multi type support is still experimental.
Sample sentence of the data:
<START:person> Pierre Vinken <END> , 61 years old , will join the board as a nonexecutive director Nov. 29 . Mr . <START:person> Vinken <END> is chairman of Elsevier N.V. , the Dutch publishing group .
The training data should contain at least 15000 sentences to create a model which performs well. Usage of the tool:
$ opennlp TokenNameFinderTrainer Usage: opennlp TokenNameFinderTrainer[.bionlp2004|.conll03|.conll02|.ad] [-resources resourcesDir] \ [-type modelType] [-featuregen featuregenFile] [-params paramsFile] \ [-iterations num] [-cutoff num] -model modelFile -lang language \ -data sampleData [-encoding charsetName] Arguments description: -resources resourcesDir The resources directory -type modelType The type of the token name finder model -featuregen featuregenFile The feature generator descriptor file -params paramsFile training parameters file. -iterations num number of training iterations, ignored if -params is used. -cutoff num minimal number of times a feature must be seen, ignored if -params is used. -model modelFile output model file. -lang language language which is being processed. -data sampleData data to be used, usually a file name. -encoding charsetName encoding for reading and writing text, if absent the system default is used.
It is now assumed that the english person name finder model should be trained from a file called en-ner-person.train which is encoded as UTF-8. The following command will train the name finder and write the model to en-ner-person.bin:
$ opennlp TokenNameFinderTrainer -model en-ner-person.bin -lang en -data en-ner-person.train -encoding UTF-8
Additionally it is possible to specify the number of iterations, the cutoff and to overwrite all types in the training data with a single type.
To train the name finder from within an application it is recommended to use the training API instead of the command line tool. Basically three steps are necessary to train it:
The application must open a sample data stream
Call the NameFinderME.train method
Save the TokenNameFinderModel to a file or database
The three steps are illustrated by the following sample code:
Charset charset = Charset.forName("UTF-8"); ObjectStream<String> lineStream = new PlainTextByLineStream(new FileInputStream("en-ner-person.train"), charset); ObjectStream<NameSample> sampleStream = new NameSampleDataStream(lineStream); TokenNameFinderModel model; try { model = NameFinderME.train("en", "person", sampleStream, TrainingParameters.defaultParams(), null, Collections.<String, Object>emptyMap()); } finally { sampleStream.close(); } try { modelOut = new BufferedOutputStream(new FileOutputStream(modelFile)); model.serialize(modelOut); } finally { if (modelOut != null) modelOut.close(); }
OpenNLP defines a default feature generation which is used when no custom feature generation is specified. Users which want to experiment with the feature generation can provide a custom feature generator. Either via API or via an xml descriptor file.
The custom generator must be used for training and for detecting the names. If the feature generation during training time and detection time is different the name finder might not be able to detect names. The following lines show how to construct a custom feature generator
AdaptiveFeatureGenerator featureGenerator = new CachedFeatureGenerator( new AdaptiveFeatureGenerator[]{ new WindowFeatureGenerator(new TokenFeatureGenerator(), 2, 2), new WindowFeatureGenerator(new TokenClassFeatureGenerator(true), 2, 2), new OutcomePriorFeatureGenerator(), new PreviousMapFeatureGenerator(), new BigramNameFeatureGenerator(), new SentenceFeatureGenerator(true, false) });
which is similar to the default feature generator. The javadoc of the feature generator classes explain what the individual feature generators do. To write a custom feature generator please implement the AdaptiveFeatureGenerator interface or if it must not be adaptive extend the FeatureGeneratorAdapter. The train method which should be used is defined as
public static TokenNameFinderModel train(String languageCode, String type, ObjectStream<NameSample> samples, TrainingParameters trainParams, AdaptiveFeatureGenerator generator, final Map<String, Object> resources) throws IOException
and can take feature generator as an argument. To detect names the model which was returned from the train method and the feature generator must be passed to the NameFinderME constructor.
new NameFinderME(model, featureGenerator, NameFinderME.DEFAULT_BEAM_SIZE);
OpenNLP can also use a xml descriptor file to configure the feature generation. The descriptor file is stored inside the model after training and the feature generators are configured correctly when the name finder is instantiated. The following sample shows a xml descriptor:
<generators> <cache> <generators> <window prevLength = "2" nextLength = "2"> <tokenclass/> </window> <window prevLength = "2" nextLength = "2"> <token/> </window> <definition/> <prevmap/> <bigram/> <sentence begin="true" end="false"/> </generators> </cache> </generators>
The root element must be generators, each sub-element adds a feature generator to the configuration. The sample xml is equivalent to the generators defined by the API above.
The following table shows the supported elements:
Table 4.1. Generator elements
Element | Aggregated | Attributes |
---|---|---|
generators | yes | none |
cache | yes | none |
charngram | no | min and max specify the length of the generated character ngrams |
definition | no | none |
dictionary | no | dict is the key of the dictionary resource to use, and prefix is a feature prefix string |
prevmap | no | none |
sentence | no | begin and end to generate begin or end features, both are optional and are boolean values |
tokenclass | no | none |
token | no | none |
bigram | no | none |
tokenpattern | no | none |
window | yes | prevLength and nextLength must be integers ans specify the window size |
custom | no | class is the name of the feature generator class which will be loaded |
Aggregated feature generators can contain other generators, like the cache or the window feature
generator in the sample.
The built in evaluation can measure the named entity recognition performance of the name finder. The performance is either measured on a test dataset or via cross validation.
The following command shows how the tool can be run:
$ opennlp TokenNameFinderEvaluator -model en-ner-person.bin -lang en -data en-ner-person.test -encoding UTF-8 Precision: 0.8005071889818507 Recall: 0.7450581122145297 F-Measure: 0.7717879983140168
Note: The command line interface does not support cross evaluation in the current version.
The evaluation can be performed on a pre-trained model and a test dataset or via cross validation. In the first case the model must be loaded and a NameSample ObjectStream must be created (see code samples above), assuming these two objects exist the following code shows how to perform the evaluation:
TokenNameFinderEvaluator evaluator = new TokenNameFinderEvaluator(new NameFinderME(model)); evaluator.evaluate(sampleStream); FMeasure result = evaluator.getFMeasure(); System.out.println(result.toString());
In the cross validation case all the training arguments must be provided (see the Training API section above). To perform cross validation the ObjectStream must be resettable.
FileInputStream sampleDataIn = new FileInputStream("en-ner-person.train"); ObjectStream<NameSample> sampleStream = new PlainTextByLineStream(sampleDataIn.getChannel(), "UTF-8"); TokenNameFinderCrossValidator evaluator = new TokenNameFinderCrossValidator("en", 100, 5); evaluator.evaluate(sampleStream, 10); FMeasure result = evaluator.getFMeasure(); System.out.println(result.toString());
Annotation guidelines define what should be labeled as an entity. To build a private corpus it is important to know these guidelines and maybe write a custom one. Here is a list of publicly available annotation guidelines:
Table of Contents
The OpenNLP Document Categorizer can classify text into pre-defined categories. It is based on maximum entropy framework. For someone interested in Gross Margin, the sample text given below could be classified as GMDecrease
Major acquisitions that have a lower gross margin than the existing network also had a negative impact on the overall gross margin, but it should improve following the implementation of its integration strategies.
and the text below could be classified as GMIncrease
The upward movement of gross margin resulted from amounts pursuant to adjustments to obligations towards dealers.
To be able to classify a text, the document categorizer needs a model. The classifications are requirements-specific and hence there is no pre-built model for document categorizer under OpenNLP project.
The easiest way to try out the document categorizer is the command line tool. The tool is only intended for demonstration and testing. The following command shows how to use the document categorizer tool.
$ opennlp Doccat model
The input is read from standard input and output is written to standard output, unless they are redirected or piped. As with most components in OpenNLP, document categorizer expects input which is segmented into sentences.
To perform classification you will need a maxent model - these are encapsulated in the DoccatModel class of OpenNLP tools.
First you need to grab the bytes from the serialized model on an InputStream - we'll leave it you to do that, since you were the one who serialized it to begin with. Now for the easy part:
InputStream is = ... DoccatModel m = new DoccatModel(is);
With the DoccatModel in hand we are just about there:
String inputText = ... DocumentCategorizerME myCategorizer = new DocumentCategorierME(m); double[] outcomes = myCategorizer.categorize(inputText); String category = myCategorizer.getBestOutcome();
The Document Categorizer can be trained on annotated training material. The data can be in OpenNLP Document Categorizer training format. This is one document per line, containing category and text separated by a whitespace. Other formats can also be available. The following sample shows the sample from above in the required format. Here GMDecrease and GMIncrease are the categories.
GMDecrease Major acquisitions that have a lower gross margin than the existing network also \ had a negative impact on the overall gross margin, but it should improve following \ the implementation of its integration strategies . GMIncrease The upward movement of gross margin resulted from amounts pursuant to adjustments \ to obligations towards dealers .
Note: The line breaks marked with a backslash are just inserted for formatting purposes and must not be included in the training data.
The following command will train the document categorizer and write the model to en-doccat.bin:
$ opennlp DoccatTrainer -model en-doccat.bin -lang en -data en-doccat.train -encoding UTF-8
Additionally it is possible to specify the number of iterations, and the cutoff.
So, naturally you will need some access to many pre-classified events to train your model. The class opennlp.tools.doccat.DocumentSample encapsulates a text document and its classification. DocumentSample has two constructors. Each take the text's category as one argument. The other argument can either be raw text, or an array of tokens. By default, the raw text will be split into tokens by whitespace. So, let's say your training data was contained in a text file, where the format is as described above. Then you might want to write something like this to create a collection of DocumentSamples:
DoccatModel model = null; InputStream dataIn = null; try { dataIn = new FileInputStream("en-sentiment.train"); ObjectStream<String> lineStream = new PlainTextByLineStream(dataIn, "UTF-8"); ObjectStream<DocumentSample> sampleStream = new DocumentSampleStream(lineStream); model = DocumentCategorizerME.train("en", sampleStream); } catch (IOException e) { // Failed to read or parse training data, training failed e.printStackTrace(); } finally { if (dataIn != null) { try { dataIn.close(); } catch (IOException e) { // Not an issue, training already finished. // The exception should be logged and investigated // if part of a production system. e.printStackTrace(); } } }
Now might be a good time to cruise over to Hulu or something, because this could take a while if you've got a large training set. You may see a lot of output as well. Once you're done, you can pretty quickly step to classification directly, but first we'll cover serialization. Feel free to skim.
OutputStream modelOut = null; try { modelOut = new BufferedOutputStream(new FileOutputStream(modelFile)); model.serialize(modelOut); } catch (IOException e) { // Failed to save model e.printStackTrace(); } finally { if (modelOut != null) { try { modelOut.close(); } catch (IOException e) { // Failed to correctly save model. // Written model might be invalid. e.printStackTrace(); } } }
Table of Contents
The Part of Speech Tagger marks tokens with their corresponding word type based on the token itself and the context of the token. A token might have multiple pos tags depending on the token and the context. The OpenNLP POS Tagger uses a probability model to predict the correct pos tag out of the tag set. To limit the possible tags for a token a tag dictionary can be used which increases the tagging and runtime performance of the tagger.
The easiest way to try out the POS Tagger is the command line tool. The tool is only intended for demonstration and testing. Download the english maxent pos model and start the POS Tagger Tool with this command:
$ opennlp POSTagger en-pos-maxent.bin
The POS Tagger now reads a tokenized sentence per line from stdin. Copy these two sentences to the console:
Pierre Vinken , 61 years old , will join the board as a nonexecutive director Nov. 29 . Mr. Vinken is chairman of Elsevier N.V. , the Dutch publishing group .
the POS Tagger will now echo the sentences with pos tags to the console:
Pierre_NNP Vinken_NNP ,_, 61_CD years_NNS old_JJ ,_, will_MD join_VB the_DT board_NN as_IN a_DT nonexecutive_JJ director_NN Nov._NNP 29_CD ._. Mr._NNP Vinken_NNP is_VBZ chairman_NN of_IN Elsevier_NNP N.V._NNP ,_, the_DT Dutch_NNP publishing_VBG group_NN
The tag set used by the english pos model is the Penn Treebank tag set.
The POS Tagger can be embedded into an application via its API. First the pos model must be loaded into memory from disk or an other source. In the sample below its loaded from disk.
InputStream modelIn = null; try { modelIn = new FileInputStream("en-pos-maxent.bin"); POSModel model = new POSModel(modelIn); } catch (IOException e) { // Model loading failed, handle the error e.printStackTrace(); } finally { if (modelIn != null) { try { modelIn.close(); } catch (IOException e) { } } }
After the model is loaded the POSTaggerME can be instantiated.
POSTaggerME tagger = new POSTaggerME(model);
The POS Tagger instance is now ready to tag data. It expects a tokenized sentence as input, which is represented as a String array, each String object in the array is one token.
The following code shows how to determine the most likely pos tag sequence for a sentence.
String sent[] = new String[]{"Most", "large", "cities", "in", "the", "US", "had", "morning", "and", "afternoon", "newspapers", "."}; String tags[] = tagger.tag(sent);
The tags array contains one part-of-speech tag for each token in the input array. The corresponding tag can be found at the same index as the token has in the input array. The confidence scores for the returned tags can be easily retrieved from a POSTaggerME with the following method call:
double probs[] = tagger.probs();
The call to probs is stateful and will always return the probabilities of the last tagged sentence. The probs method should only be called when the tag method was called before, otherwise the behavior is undefined.
Some applications need to retrieve the n-best pos tag sequences and not only the best sequence. The topKSequences method is capable of returning the top sequences. It can be called in a similar way as tag.
Sequence topSequences[] = tagger.topKSequences(sent);
Each Sequence object contains one sequence. The sequence can be retrieved via Sequence.getOutcomes() which returns a tags array and Sequence.getProbs() returns the probability array for this sequence.
The POS Tagger can be trained on annotated training material. The training material is a collection of tokenized sentences where each token has the assigned part-of-speech tag. The native POS Tagger training material looks like this:
About_IN 10_CD Euro_NNP ,_, I_PRP reckon_VBP ._. That_DT sounds_VBZ good_JJ ._.
Each sentence must be in one line. The token/tag pairs are combined with "_". The token/tag pairs are whitespace separated. The data format does not define a document boundary. If a document boundary should be included in the training material it is suggested to use an empty line.
The Part-of-Speech Tagger can either be trained with a command line tool, or via an training API.
OpenNLP has a command line tool which is used to train the models available from the model download page on various corpora.
Usage of the tool:
$ opennlp POSTaggerTrainer Usage: opennlp POSTaggerTrainer[.conllx] [-type maxent|perceptron|perceptron_sequence] \ [-dict dictionaryPath] [-ngram cutoff] [-params paramsFile] [-iterations num] \ [-cutoff num] -model modelFile -lang language -data sampleData \ [-encoding charsetName] Arguments description: -type maxent|perceptron|perceptron_sequence The type of the token name finder model. One of maxent|perceptron|perceptron_sequence. -dict dictionaryPath The XML tag dictionary file -ngram cutoff NGram cutoff. If not specified will not create ngram dictionary. -params paramsFile training parameters file. -iterations num number of training iterations, ignored if -params is used. -cutoff num minimal number of times a feature must be seen, ignored if -params is used. -model modelFile output model file. -lang language language which is being processed. -data sampleData data to be used, usually a file name. -encoding charsetName encoding for reading and writing text, if absent the system default is used.
The following command illustrates how an english part-of-speech model can be trained:
$ opennlp POSTaggerTrainer -type maxent -model en-pos-maxent.bin \ -lang en -data en-pos.train -encoding UTF-8
The Part-of-Speech Tagger training API supports the training of a new pos model. Basically three steps are necessary to train it:
The application must open a sample data stream
Call the POSTagger.train method
Save the POSModel to a file or database
The following code illustrates that:
POSModel model = null; InputStream dataIn = null; try { dataIn = new FileInputStream("en-pos.train"); ObjectStream<String> lineStream = new PlainTextByLineStream(dataIn, "UTF-8"); ObjectStream<POSSample> sampleStream = new WordTagSampleStream(lineStream); model = POSTaggerME.train("en", sampleStream, TrainingParameters.defaultParams(), null, null); } catch (IOException e) { // Failed to read or parse training data, training failed e.printStackTrace(); } finally { if (dataIn != null) { try { dataIn.close(); } catch (IOException e) { // Not an issue, training already finished. // The exception should be logged and investigated // if part of a production system. e.printStackTrace(); } } }
The above code performs the first two steps, opening the data and training the model. The trained model must still be saved into an OutputStream, in the sample below it is written into a file.
OutputStream modelOut = null; try { modelOut = new BufferedOutputStream(new FileOutputStream(modelFile)); model.serialize(modelOut); } catch (IOException e) { // Failed to save model e.printStackTrace(); } finally { if (modelOut != null) { try { modelOut.close(); } catch (IOException e) { // Failed to correctly save model. // Written model might be invalid. e.printStackTrace(); } }
The tag dictionary is a word dictionary which specifies which tags a specific token can have. Using a tag dictionary has two advantages, inappropriate tags can not been assigned to tokens in the dictionary and the beam search algorithm has to consider less possibilities and can search faster.
The dictionary is defined in a xml format and can be created and stored with the POSDictionary class. Please for now checkout the javadoc and source code of that class.
Note: The format should be documented and sample code should show how to use the dictionary. Any contributions are very welcome. If you want to contribute please contact us on the mailing list or comment on the jira issue OPENNLP-287.
The built in evaluation can measure the accuracy of the pos tagger. The accuracy can be measured on a test data set or via cross validation.
There is a command line tool to evaluate a given model on a test data set. The following command shows how the tool can be run:
$ opennlp POSTaggerEvaluator -model pt.postagger.bin -lang pt -data pt.postagger.test -encoding utf-8
This will display the resulting accuracy score, e.g.:
Loading model ... done Evaluating ... done Accuracy: 0.9659110277825124
There is a command line tool to cross validate a test data set. The following command shows how the tool can be run:
$ opennlp POSTaggerCrossValidator -lang pt -data pt.postagger.test -encoding utf-8
This will display the resulting accuracy score, e.g.:
Accuracy: 0.9659110277825124
Table of Contents
Text chunking consists of dividing a text in syntactically correlated parts of words, like noun groups, verb groups, but does not specify their internal structure, nor their role in the main sentence.
The easiest way to try out the Chunker is the command line tool. The tool is only intended for demonstration and testing.
Download the english maxent chunker model from the website and start the Chunker Tool with this command:
$ opennlp ChunkerME en-chunker.bin
The Chunker now reads a pos tagged sentence per line from stdin. Copy these two sentences to the console:
Rockwell_NNP International_NNP Corp._NNP 's_POS Tulsa_NNP unit_NN said_VBD it_PRP signed_VBD a_DT tentative_JJ agreement_NN extending_VBG its_PRP$ contract_NN with_IN Boeing_NNP Co._NNP to_TO provide_VB structural_JJ parts_NNS for_IN Boeing_NNP 's_POS 747_CD jetliners_NNS ._. Rockwell_NNP said_VBD the_DT agreement_NN calls_VBZ for_IN it_PRP to_TO supply_VB 200_CD additional_JJ so-called_JJ shipsets_NNS for_IN the_DT planes_NNS ._.
The Chunker will now echo the sentences grouped tokens to the console:
[NP Rockwell_NNP International_NNP Corp._NNP ] [NP 's_POS Tulsa_NNP unit_NN ] [VP said_VBD ] [NP it_PRP ] [VP signed_VBD ] [NP a_DT tentative_JJ agreement_NN ] [VP extending_VBG ] [NP its_PRP$ contract_NN ] [PP with_IN ] [NP Boeing_NNP Co._NNP ] [VP to_TO provide_VB ] [NP structural_JJ parts_NNS ] [PP for_IN ] [NP Boeing_NNP ] [NP 's_POS 747_CD jetliners_NNS ] ._. [NP Rockwell_NNP ] [VP said_VBD ] [NP the_DT agreement_NN ] [VP calls_VBZ ] [SBAR for_IN ] [NP it_PRP ] [VP to_TO supply_VB ] [NP 200_CD additional_JJ so-called_JJ shipsets_NNS ] [PP for_IN ] [NP the_DT planes_NNS ] ._.
The tag set used by the english pos model is the Penn Treebank tag set.
The Chunker can be embedded into an application via its API. First the chunker model must be loaded into memory from disk or an other source. In the sample below its loaded from disk.
InputStream modelIn = null; ChunkerModel model = null; try { modelIn = new FileInputStream("en-chunker.bin"); model = new ChunkerModel(modelIn); } catch (IOException e) { // Model loading failed, handle the error e.printStackTrace(); } finally { if (modelIn != null) { try { modelIn.close(); } catch (IOException e) { } } }
After the model is loaded a Chunker can be instantiated.
ChunkerME chunker = new ChunkerME(model);
The Chunker instance is now ready to tag data. It expects a tokenized sentence as input, which is represented as a String array, each String object in the array is one token, and the POS tags associated with each token.
The following code shows how to determine the most likely chunk tag sequence for a sentence.
String sent[] = new String[] { "Rockwell", "International", "Corp.", "'s", "Tulsa", "unit", "said", "it", "signed", "a", "tentative", "agreement", "extending", "its", "contract", "with", "Boeing", "Co.", "to", "provide", "structural", "parts", "for", "Boeing", "'s", "747", "jetliners", "." }; String pos[] = new String[] { "NNP", "NNP", "NNP", "POS", "NNP", "NN", "VBD", "PRP", "VBD", "DT", "JJ", "NN", "VBG", "PRP$", "NN", "IN", "NNP", "NNP", "TO", "VB", "JJ", "NNS", "IN", "NNP", "POS", "CD", "NNS", "." }; String tag[] = chunker.chunk(sent, pos);
The tags array contains one chunk tag for each token in the input array. The corresponding tag can be found at the same index as the token has in the input array. The confidence scores for the returned tags can be easily retrieved from a ChunkerME with the following method call:
double probs[] = chunker.probs();
The call to probs is stateful and will always return the probabilities of the last tagged sentence. The probs method should only be called when the tag method was called before, otherwise the behavior is undefined.
Some applications need to retrieve the n-best chunk tag sequences and not only the best sequence. The topKSequences method is capable of returning the top sequences. It can be called in a similar way as chunk.
Sequence topSequences[] = chunk.topKSequences(sent, pos);
Each Sequence object contains one sequence. The sequence can be retrieved via Sequence.getOutcomes() which returns a tags array and Sequence.getProbs() returns the probability array for this sequence.
The pre-trained models might not be available for a desired language, can not detect important entities or the performance is not good enough outside the news domain.
These are the typical reason to do custom training of the chunker on a ne corpus or on a corpus which is extended by private training data taken from the data which should be analyzed.
The training data can be converted to the OpenNLP chunker training format, that is based on CoNLL2000. Other formats may also be available. The train data consist of three columns separated by spaces. Each word has been put on a separate line and there is an empty line after each sentence. The first column contains the current word, the second its part-of-speech tag and the third its chunk tag. The chunk tags contain the name of the chunk type, for example I-NP for noun phrase words and I-VP for verb phrase words. Most chunk types have two types of chunk tags, B-CHUNK for the first word of the chunk and I-CHUNK for each other word in the chunk. Here is an example of the file format:
Sample sentence of the training data:
He PRP B-NP reckons VBZ B-VP the DT B-NP current JJ I-NP account NN I-NP deficit NN I-NP will MD B-VP narrow VB I-VP to TO B-PP only RB B-NP # # I-NP 1.8 CD I-NP billion CD I-NP in IN B-PP September NNP B-NP . . O
OpenNLP has a command line tool which is used to train the models available from the model download page on various corpora.
Usage of the tool:
$ opennlp ChunkerTrainerME Usage: opennlp ChunkerTrainerME[.ad] [-params paramsFile] [-iterations num] [-cutoff num] \ -model modelFile -lang language -data sampleData [-encoding charsetName] Arguments description: -params paramsFile training parameters file. -iterations num number of training iterations, ignored if -params is used. -cutoff num minimal number of times a feature must be seen, ignored if -params is used. -model modelFile output model file. -lang language language which is being processed. -data sampleData data to be used, usually a file name. -encoding charsetName encoding for reading and writing text, if absent the system default is used.
Its now assumed that the english chunker model should be trained from a file called en-chunker.train which is encoded as UTF-8. The following command will train the name finder and write the model to en-chunker.bin:
$ opennlp ChunkerTrainerME -model en-chunker.bin -lang en -data en-chunker.train -encoding UTF-8
Additionally its possible to specify the number of iterations, the cutoff and to overwrite all types in the training data with a single type.
The Chunker offers an API to train a new chunker model. The following sample code illustrates how to do it:
Charset charset = Charset.forName("UTF-8"); ObjectStream<String> lineStream = new PlainTextByLineStream(new FileInputStream("en-chunker.train"),charset); ObjectStream<ChunkSample> sampleStream = new ChunkSampleStream(lineStream); ChunkerModel model; try { model = ChunkerME.train("en", sampleStream, new DefaultChunkerContextGenerator(), TrainingParameters.defaultParams()); } finally { sampleStream.close(); } OutputStream modelOut = null; try { modelOut = new BufferedOutputStream(new FileOutputStream(modelFile)); model.serialize(modelOut); } finally { if (modelOut != null) modelOut.close(); }
The built in evaluation can measure the chunker performance. The performance is either measured on a test dataset or via cross validation.
The following command shows how the tool can be run:
$ opennlp ChunkerEvaluator Usage: opennlp ChunkerEvaluator[.ad] -model model [-misclassified true|false] \ [-detailedF true|false] -lang language -data sampleData [-encoding charsetName]
A sample of the command considering you have a data sample named en-chunker.eval and you trained a model called en-chunker.bin:
$ opennlp ChunkerEvaluator -model en-chunker.bin -lang en -data en-chunker.eval -encoding UTF-8
and here is a sample output:
Precision: 0.9255923572240226 Recall: 0.9220610430991112 F-Measure: 0.9238233255623465
You can also use the tool to perform 10-fold cross validation of the Chunker. he following command shows how the tool can be run:
$ opennlp ChunkerCrossValidator Usage: opennlp ChunkerCrossValidator[.ad] [-params paramsFile] [-iterations num] [-cutoff num] \ [-misclassified true|false] [-folds num] [-detailedF true|false] \ -lang language -data sampleData [-encoding charsetName] Arguments description: -params paramsFile training parameters file. -iterations num number of training iterations, ignored if -params is used. -cutoff num minimal number of times a feature must be seen, ignored if -params is used. -misclassified true|false if true will print false negatives and false positives. -folds num number of folds, default is 10. -detailedF true|false if true will print detailed FMeasure results. -lang language language which is being processed. -data sampleData data to be used, usually a file name. -encoding charsetName encoding for reading and writing text, if absent the system default is used.
It is not necessary to pass a model. The tool will automatically split the data to train and evaluate:
$ opennlp ChunkerCrossValidator -lang pt -data en-chunker.cross -encoding UTF-8
Table of Contents
The easiest way to try out the Parser is the command line tool. The tool is only intended for demonstration and testing. Download the English chunking parser model from the our website and start the Parse Tool with the following command.
$ opennlp Parser en-parser.bin en-parser-chunking.bin
Loading the big parser model can take several seconds, be patient. Copy this sample sentence to the console.
The quick brown fox jumps over the lazy dog .
The parser should now print the following to the console.
(TOP (NP (NP (DT The) (JJ quick) (JJ brown) (NN fox) (NNS jumps)) (PP (IN over) (NP (DT the) (JJ lazy) (NN dog))) (. .)))
With the following command the input can be read from a file and be written to an output file.
$ opennlp Parser en-parser.bin en-parser-chunking.bin < article-tokenized.txt > article-parsed.txt.
The article-tokenized.txt file must contain one sentence per line which is tokenized with the English tokenizer model from our website. See the Tokenizer documentation for further details.
The Parser can be easily integrated into an application via its API. To instantiate a Parser the parser model must be loaded first.
InputStream modelIn = new FileInputStream("en-parser-chunking.bin"); try { ParserModel model = new ParserModel(modelIn); } catch (IOException e) { e.printStackTrace(); } finally { if (modelIn != null) { try { modelIn.close(); } catch (IOException e) { } } }
Unlike the other components to instantiate the Parser a factory method should be used instead of creating the Parser via the new operator. The parser model is either trained for the chunking parser or the tree insert parser the parser implementation must be chosen correctly. The factory method will read a type parameter from the model and create an instance of the corresponding parser implementation.
Parser parser = ParserFactory.create(model);
Right now the tree insert parser is still experimental and there is no pre-trained model for it. The parser expect a whitespace tokenized sentence. A utility method from the command line tool can parse the sentence String. The following code shows how the parser can be called.
String sentence = "The quick brown fox jumps over the lazy dog .";
Parse topParses[] = ParserTool.parseLine(sentence, parser, 1);
The topParses array only contains one parse because the number of parses is set to 1. The Parse object contains the parse tree. To display the parse tree call the show method. It either prints the parse to the console or into a provided StringBuffer. Similar to Exception.printStackTrace.
TODO: Extend this section with more information about the Parse object.
The OpenNLP offers two different parser implementations, the chunking parser and the treeinsert parser. The later one is still experimental and not recommended for production use. (TODO: Add a section which explains the two different approaches) The training can either be done with the command line tool or the training API. In the first case the training data must be available in the OpenNLP format. Which is the Penn Treebank format, but with the limitation of a sentence per line.
(TOP (S (NP-SBJ (DT Some) )(VP (VBP say) (NP (NNP November) ))(. .) )) (TOP (S (NP-SBJ (PRP I) )(VP (VBP say) (NP (CD 1992) ))(. .) ('' '') ))
Penn Treebank annotation guidelines can be found on the Penn Treebank home page. A parser model also contains a pos tagger model, depending on the amount of available training data it is recommend to switch the tagger model against a tagger model which was trained on a larger corpus. The pre-trained parser model provided on the website is doing this to achieve a better performance. (TODO: On which data is the model on the website trained, and say on which data the tagger model is trained)
OpenNLP has a command line tool which is used to train the models available from the model download page on various corpora. The data must be converted to the OpenNLP parser training format, which is shortly explained above. To train the parser a head rules file is also needed. (TODO: Add documentation about the head rules file) Usage of the tool:
$ opennlp ParserTrainer Usage: opennlp ParserTrainer -headRules headRulesFile [-parserType CHUNKING|TREEINSERT] \ [-params paramsFile] [-iterations num] [-cutoff num] \ -model modelFile -lang language -data sampleData \ [-encoding charsetName] Arguments description: -headRules headRulesFile head rules file. -parserType CHUNKING|TREEINSERT one of CHUNKING or TREEINSERT, default is CHUNKING. -params paramsFile training parameters file. -iterations num number of training iterations, ignored if -params is used. -cutoff num minimal number of times a feature must be seen, ignored if -params is used. -model modelFile output model file. -format formatName data format, might have its own parameters. -encoding charsetName encoding for reading and writing text, if absent the system default is used. -lang language language which is being processed. -data sampleData data to be used, usually a file name. -encoding charsetName encoding for reading and writing text, if absent the system default is used.
The model on the website was trained with the following command:
$ opennlp ParserTrainer -model en-parser-chunking.bin -parserType CHUNKING \ -head-rules head_rules \ -lang en -data train.all -encoding ISO-8859-1
Its also possible to specify the cutoff and the number of iterations, these parameters are used for all trained models. The -parserType parameter is an optional parameter, to use the tree insertion parser, specify TREEINSERT as type. The TaggerModelReplacer tool replaces the tagger model inside the parser model with a new one.
Note: The original parser model will be overwritten with the new parser model which contains the replaced tagger model.
$ opennlp TaggerModelReplacer en-parser-chunking.bin en-pos-maxent.bin
Additionally there are tools to just retrain the build or the check model.
TODO: Write documentation about the parser training api. Any contributions are very welcome. If you want to contribute please contact us on the mailing list or comment on the jira issue OPENNLP-219.
The built in evaluation can measure the parser performance. The performance is measured on a test dataset.
The following command shows how the tool can be run:
$ opennlp ParserEvaluator Usage: opennlp ParserEvaluator[.ontonotes|frenchtreebank] [-misclassified true|false] -model model \ -data sampleData [-encoding charsetName]
A sample of the command considering you have a data sample named en-parser-chunking.eval and you trained a model called en-parser-chunking.bin:
$ opennlp ParserEvaluator -model en-parser-chunking.bin -lang en -data en-parser-chunking.eval -encoding UTF-8
and here is a sample output:
Precision: 0.9009744742967609 Recall: 0.8962012400910446 F-Measure: 0.8985815184245214
The Parser Evaluation tool reimplements the PARSEVAL scoring method as implemented by the EVALB script, which is the most widely used evaluation tool for constituent parsing. Note however that currently the Parser Evaluation tool does not allow to make exceptions in the constituents to be evaluated, in the way Collins or Bikel usually do. Any contributions are very welcome. If you want to contribute please contact us on the mailing list or comment on the jira issue OPENNLP-688.
The OpenNLP Coreference Resolution system links multiple mentions of an entity in a document together. The OpenNLP implementation is currently limited to noun phrase mentions, other mention types cannot be resolved.
TODO: Write more documentation about the coref component. Any contributions are very welcome. If you want to contribute please contact us on the mailing list or comment on the jira issue OPENNLP-48.
Table of Contents
In OpenNLP extension can be used to add new functionality and to heavily customize an existing component. Most components define a factory class which can be implemented to customize the creation of it. And some components allow to add new feature generators.
In many places it is possible to pass in an extension class name to customize some aspect of OpenNLP. The implementation class needs to implement the specified interface and should have a public no-argument constructor.
The traditional way of loading an extension via Class.forName does not work in an OSGi environment because the class paths of the OpenNLP Tools and extension bundle are isolated. OSGi uses services to provide functionality from one bundle to another. The extension bundle must register its extensions as services so that the OpenNLP tools bundle can use them. The following code illustrates how that can be done:
Dictionary<String, String> props = new Hashtable<String, String>(); props.put(ExtensionServiceKeys.ID, "org.test.SuperTokenizer"); context.registerService(Tokenizer.class.getName(), new org.test.SuperTokenizer(), props);
The service OpenNLP is looking for might not be (yet) available. In this case OpenNLP waits until a timeout is reached. If loading the extension fails an ExtensionNotLoadedException is thrown. This exception is also thrown when the thread is interrupted while it is waiting for the extension, the interrupted flag will be set again and the calling code has a chance to handle it.
Table of Contents
OpenNLP has built-in support to convert into the native training format or directly use various corpora needed by the different trainable components.
CoNLL stands for the Conference on Computational Natural Language Learning and is not a single project but a consortium of developers attempting to broaden the computing environment. More information about the entire conference series can be obtained here for CoNLL.
The shared task of CoNLL-2000 is Chunking.
CoNLL-2000 made available training and test data for the Chunk task in English. The data consists of the same partitions of the Wall Street Journal corpus (WSJ) as the widely used data for noun phrase chunking: sections 15-18 as training data (211727 tokens) and section 20 as test data (47377 tokens). The annotation of the data has been derived from the WSJ corpus by a program written by Sabine Buchholz from Tilburg University, The Netherlands. Both training and test data can be obtained from http://www.cnts.ua.ac.be/conll2000/chunking.
The data don't need to be transformed because Apache OpenNLP Chunker follows the CONLL 2000 format for training. Check Chunker Training section to learn more.
We can train the model for the Chunker using the train.txt available at CONLL 2000:
$ opennlp ChunkerTrainerME -model en-chunker.bin -iterations 500 \ -lang en -data train.txt -encoding UTF-8
Indexing events using cutoff of 5 Computing event counts... done. 211727 events Indexing... done. Sorting and merging events... done. Reduced 211727 events to 197252. Done indexing. Incorporating indexed data for training... done. Number of Event Tokens: 197252 Number of Outcomes: 22 Number of Predicates: 107838 ...done. Computing model parameters... Performing 500 iterations. 1: .. loglikelihood=-654457.1455212828 0.2601510435608118 2: .. loglikelihood=-239513.5583724216 0.9260037690044255 3: .. loglikelihood=-141313.1386347238 0.9443387003074715 4: .. loglikelihood=-101083.50853437989 0.954375209585929 ... cut lots of iterations ... 498: .. loglikelihood=-1710.8874647317095 0.9995040783650645 499: .. loglikelihood=-1708.0908900815848 0.9995040783650645 500: .. loglikelihood=-1705.3045902366732 0.9995040783650645 Writing chunker model ... done (4.019s) Wrote chunker model to path: .\en-chunker.bin
We evaluate the model using the file test.txt available at CONLL 2000:
$ opennlp ChunkerEvaluator -model en-chunker.bin -lang en -encoding utf8 -data test.txt
Loading Chunker model ... done (0,665s) current: 85,8 sent/s avg: 85,8 sent/s total: 86 sent current: 88,1 sent/s avg: 87,0 sent/s total: 174 sent current: 156,2 sent/s avg: 110,0 sent/s total: 330 sent current: 192,2 sent/s avg: 130,5 sent/s total: 522 sent current: 167,2 sent/s avg: 137,8 sent/s total: 689 sent current: 179,2 sent/s avg: 144,6 sent/s total: 868 sent current: 183,2 sent/s avg: 150,3 sent/s total: 1052 sent current: 183,2 sent/s avg: 154,4 sent/s total: 1235 sent current: 169,2 sent/s avg: 156,0 sent/s total: 1404 sent current: 178,2 sent/s avg: 158,2 sent/s total: 1582 sent current: 172,2 sent/s avg: 159,4 sent/s total: 1754 sent current: 177,2 sent/s avg: 160,9 sent/s total: 1931 sent Average: 161,6 sent/s Total: 2013 sent Runtime: 12.457s Precision: 0.9244354736974896 Recall: 0.9216837162502096 F-Measure: 0.9230575441395671
The shared task of CoNLL-2002 is language independent named entity recognition for Spanish and Dutch.
The data consists of three files per language: one training file and two test files testa and testb. The first test file will be used in the development phase for finding good parameters for the learning system. The second test file will be used for the final evaluation. Currently there are data files available for two languages: Spanish and Dutch.
The Spanish data is a collection of news wire articles made available by the Spanish EFE News Agency. The articles are from May 2000. The annotation was carried out by the TALP Research Center of the Technical University of Catalonia (UPC) and the Center of Language and Computation (CLiC)of the University of Barcelona (UB), and funded by the European Commission through the NAMIC project (IST-1999-12392).
The Dutch data consist of four editions of the Belgian newspaper "De Morgen" of 2000 (June 2, July 1, August 1 and September 1). The data was annotated as a part of the Atranos project at the University of Antwerp.
You can find the Spanish files here: http://www.lsi.upc.edu/~nlp/tools/nerc/nerc.html You must download esp.train.gz, unzip it and you will see the file esp.train.
You can find the Dutch files here: http://www.cnts.ua.ac.be/conll2002/ner.tgz You must unzip it and go to /ner/data/ned.train.gz, so you unzip it too, and you will see the file ned.train.
I will use Spanish data as reference, but it would be the same operations to Dutch. You just must remember change “-lang es” to “-lang nl” and use the correct training files. So to convert the information to the OpenNLP format:
$ opennlp TokenNameFinderConverter conll02 -data esp.train -lang es -types per > es_corpus_train_persons.txt
Optionally, you can convert the training test samples as well.
$ opennlp TokenNameFinderConverter conll02 -data esp.testa -lang es -types per > corpus_testa.txt $ opennlp TokenNameFinderConverter conll02 -data esp.testb -lang es -types per > corpus_testb.txt
To train the model for the name finder:
\bin\opennlp TokenNameFinderTrainer -lang es -encoding u tf8 -iterations 500 -data es_corpus_train_persons.txt -model es_ner_person.bin Indexing events using cutoff of 5 Computing event counts... done. 264715 events Indexing... done. Sorting and merging events... done. Reduced 264715 events to 222660. Done indexing. Incorporating indexed data for training... done. Number of Event Tokens: 222660 Number of Outcomes: 3 Number of Predicates: 71514 ...done. Computing model parameters ... Performing 500 iterations. 1: ... loglikelihood=-290819.1519958615 0.9689326256540053 2: ... loglikelihood=-37097.17676455632 0.9689326256540053 3: ... loglikelihood=-22910.372489660916 0.9706476776911017 4: ... loglikelihood=-17091.547325669497 0.9777874317662392 5: ... loglikelihood=-13797.620926769372 0.9833821279489262 6: ... loglikelihood=-11715.806710780415 0.9867140131839903 7: ... loglikelihood=-10289.222078246517 0.9886859452618855 8: ... loglikelihood=-9249.208318314624 0.9902310031543358 9: ... loglikelihood=-8454.169590899777 0.9913227433277298 10: ... loglikelihood=-7823.742997451327 0.9921953799369133 11: ... loglikelihood=-7309.375882641964 0.9928224694482746 12: ... loglikelihood=-6880.131972149693 0.9932946754056249 13: ... loglikelihood=-6515.3828767792365 0.993638441342576 14: ... loglikelihood=-6200.82723154046 0.9939595413935742 15: ... loglikelihood=-5926.213730444915 0.994269308501596 16: ... loglikelihood=-5683.9821840753275 0.9945299661900534 17: ... loglikelihood=-5468.4211798176075 0.9948246227074401 18: ... loglikelihood=-5275.127017232056 0.9950286156810154 ... cut lots of iterations ... 491: ... loglikelihood=-1174.8485558758211 0.998983812779782 492: ... loglikelihood=-1173.9971776942477 0.998983812779782 493: ... loglikelihood=-1173.1482915871768 0.998983812779782 494: ... loglikelihood=-1172.3018855781158 0.998983812779782 495: ... loglikelihood=-1171.457947774544 0.998983812779782 496: ... loglikelihood=-1170.6164663670502 0.998983812779782 497: ... loglikelihood=-1169.7774296286693 0.998983812779782 498: ... loglikelihood=-1168.94082591387 0.998983812779782 499: ... loglikelihood=-1168.1066436580463 0.9989875904274408 500: ... loglikelihood=-1167.2748713765225 0.9989875904274408 Writing name finder model ... done (2,168s) Wrote name finder model to path: .\es_ner_person.bin
The shared task of CoNLL-2003 is language independent named entity recognition for English and German.
The English data is the Reuters Corpus, which is a collection of news wire articles. The Reuters Corpus can be obtained free of charges from the NIST for research purposes: http://trec.nist.gov/data/reuters/reuters.html
The German data is a collection of articles from the German newspaper Frankfurter Rundschau. The articles are part of the ECI Multilingual Text Corpus which can be obtained for 75$ (2010) from the Linguistic Data Consortium: http://www.ldc.upenn.edu/Catalog/CatalogEntry.jsp?catalogId=LDC94T5
After one of the corpora is available the data must be transformed as explained in the README file to the CONLL format. The transformed data can be read by the OpenNLP CONLL03 converter.
To convert the information to the OpenNLP format:
$ opennlp TokenNameFinderConverter conll03 -lang en -types per -data eng.train > corpus_train.txt
Optionally, you can convert the training test samples as well.
$ opennlp TokenNameFinderConverter conll03 -lang en -types per -data eng.testa > corpus_testa.txt $ opennlp TokenNameFinderConverter conll03 -lang en -types per -data eng.testb > corpus_testb.txt
You can train the model for the name finder this way:
$ opennlp TokenNameFinderTrainer.conll03 -model en_ner_person.bin -iterations 500 \ -lang en -types per -data eng.train -encoding utf8
If you have converted the data, then you can train the model for the name finder this way:
$ opennlp TokenNameFinderTrainer -model en_ner_person.bin -iterations 500 \ -lang en -data corpus_train.txt -encoding utf8
Either way you should see the following output during the training process:
Indexing events using cutoff of 5 Computing event counts... done. 203621 events Indexing... done. Sorting and merging events... done. Reduced 203621 events to 179409. Done indexing. Incorporating indexed data for training... done. Number of Event Tokens: 179409 Number of Outcomes: 3 Number of Predicates: 58814 ...done. Computing model parameters... Performing 500 iterations. 1: .. loglikelihood=-223700.5328318588 0.9453494482396216 2: .. loglikelihood=-40525.939777363084 0.9467933071736215 3: .. loglikelihood=-24893.98837874921 0.9598518816821447 4: .. loglikelihood=-18420.3379471033 0.9712996203731442 ... cut lots of iterations ... 498: .. loglikelihood=-952.8501399442295 0.9988950059178572 499: .. loglikelihood=-952.0600155746948 0.9988950059178572 500: .. loglikelihood=-951.2722802086295 0.9988950059178572 Writing name finder model ... done (1.638s) Wrote name finder model to path: .\en_ner_person.bin
You can evaluate the model for the name finder this way:
$ opennlp TokenNameFinderEvaluator.conll03 -model en_ner_person.bin \ -lang en -types per -data eng.testa -encoding utf8
If you converted the test A and B files above, you can use them to evaluate the model.
$ opennlp TokenNameFinderEvaluator -model en_ner_person.bin -lang en -data corpus_testa.txt \ -encoding utf8
Either way you should see the following output:
Loading Token Name Finder model ... done (0.359s) current: 190.2 sent/s avg: 190.2 sent/s total: 199 sent current: 648.3 sent/s avg: 415.9 sent/s total: 850 sent current: 530.1 sent/s avg: 453.6 sent/s total: 1380 sent current: 793.8 sent/s avg: 539.0 sent/s total: 2178 sent current: 705.4 sent/s avg: 571.9 sent/s total: 2882 sent Average: 569.4 sent/s Total: 3251 sent Runtime: 5.71s Precision: 0.9366247297154147 Recall: 0.739956568946797 F-Measure: 0.8267557582133971
The Portuguese corpora available at Floresta Sintá(c)tica project follow the Arvores Deitadas (AD) format. Apache OpenNLP includes tools to convert from AD format to native format.
The Corpus can be downloaded from here: http://www.linguateca.pt/floresta/corpus.html
The Name Finder models were trained using the Amazonia corpus: amazonia.ad. The Chunker models were trained using the Bosque_CF_8.0.ad.
To extract NameFinder training data from Amazonia corpus:
$ opennlp TokenNameFinderConverter ad -lang pt -encoding ISO-8859-1 -data amazonia.ad > corpus.txt
To extract Chunker training data from Bosque_CF_8.0.ad corpus:
$ opennlp ChunkerConverter ad -lang pt -data Bosque_CF_8.0.ad.txt -encoding ISO-8859-1 > bosque-chunk
To perform the evaluation the corpus was split into a training and a test part.
$ sed '1,55172d' corpus.txt > corpus_train.txt $ sed '55172,100000000d' corpus.txt > corpus_test.txt
$ opennlp TokenNameFinderTrainer -model pt-ner.bin -cutoff 20 -lang PT -data corpus_train.txt -encoding UTF-8 ... $ opennlp TokenNameFinderEvaluator -model pt-ner.bin -lang PT -data corpus_train.txt -encoding UTF-8 Precision: 0.8005071889818507 Recall: 0.7450581122145297 F-Measure: 0.7717879983140168
The Leipzig Corpora collection presents corpora in different languages. The corpora is a collection of individual sentences collected from the web and newspapers. The Corpora is available as plain text and as MySQL database tables. The OpenNLP integration can only use the plain text version.
The corpora in the different languages can be used to train a document categorizer model which can detect the document language. The individual plain text packages can be downloaded here: http://corpora.uni-leipzig.de/download.html
After all packages have been downloaded, unzip them and use the following commands to produce a training file which can be processed by the Document Categorizer:
$ opennlp DoccatConverter leipzig -lang cat -data Leipzig/cat100k/sentences.txt >> lang.train $ opennlp DoccatConverter leipzig -lang de -data Leipzig/de100k/sentences.txt >> lang.train $ opennlp DoccatConverter leipzig -lang dk -data Leipzig/dk100k/sentences.txt >> lang.train $ opennlp DoccatConverter leipzig -lang ee -data Leipzig/ee100k/sentences.txt >> lang.train $ opennlp DoccatConverter leipzig -lang en -data Leipzig/en100k/sentences.txt >> lang.train $ opennlp DoccatConverter leipzig -lang fi -data Leipzig/fi100k/sentences.txt >> lang.train $ opennlp DoccatConverter leipzig -lang fr -data Leipzig/fr100k/sentences.txt >> lang.train $ opennlp DoccatConverter leipzig -lang it -data Leipzig/it100k/sentences.txt >> lang.train $ opennlp DoccatConverter leipzig -lang jp -data Leipzig/jp100k/sentences.txt >> lang.train $ opennlp DoccatConverter leipzig -lang kr -data Leipzig/kr100k/sentences.txt >> lang.train $ opennlp DoccatConverter leipzig -lang nl -data Leipzig/nl100k/sentences.txt >> lang.train $ opennlp DoccatConverter leipzig -lang no -data Leipzig/no100k/sentences.txt >> lang.train $ opennlp DoccatConverter leipzig -lang se -data Leipzig/se100k/sentences.txt >> lang.train $ opennlp DoccatConverter leipzig -lang sorb -data Leipzig/sorb100k/sentences.txt >> lang.train $ opennlp DoccatConverter leipzig -lang tr -data Leipzig/tr100k/sentences.txt >> lang.train
Depending on your platform local it might be problematic to output characters which are not supported by that encoding, we suggest to run these command on a platform which has a unicode default encoding, e.g. Linux with UTF-8.
After the lang.train file is created the actual language detection document categorizer model can be created with the following command.
$ opennlp DoccatTrainer -model lang.model -lang x-unspecified -data lang.train -encoding MacRoman Indexing events using cutoff of 5 Computing event counts... done. 10000 events Indexing... done. Sorting and merging events... done. Reduced 10000 events to 10000. Done indexing. Incorporating indexed data for training... done. Number of Event Tokens: 10000 Number of Outcomes: 2 Number of Predicates: 42730 ...done. Computing model parameters... Performing 100 iterations. 1: .. loglikelihood=-6931.471805600547 0.5 2: .. loglikelihood=-2110.9654348555955 1.0 ... cut lots of iterations ... 99: .. loglikelihood=-0.449640418555347 1.0 100: .. loglikelihood=-0.443746359746235 1.0 Writing document categorizer model ... done (1.210s) Wrote document categorizer model to path: /Users/joern/dev/opennlp-apache/opennlp/opennlp-tools/lang.model
In the sample above the language detection model was trained to distinguish two languages, danish and english.
After the model is created it can be used to detect the two languages:
$ bin/opennlp Doccat ../lang. lang.model lang.train karkand:opennlp-tools joern$ bin/opennlp Doccat ../lang.model Loading Document Categorizer model ... done (0.289s) The American Finance Association is pleased to announce the award of ... en The American Finance Association is pleased to announce the award of .. . Danskerne skal betale for den økonomiske krise ved at blive længere på arbejdsmarkedet . dk Danskerne skal betale for den økonomiske krise ved at blive længere på arbejdsmarkedet .
"OntoNotes Release 4.0, Linguistic Data Consortium (LDC) catalog number LDC2011T03 and isbn 1-58563-574-X, was developed as part of the OntoNotes project, a collaborative effort between BBN Technologies, the University of Colorado, the University of Pennsylvania and the University of Southern Californias Information Sciences Institute. The goal of the project is to annotate a large corpus comprising various genres of text (news, conversational telephone speech, weblogs, usenet newsgroups, broadcast, talk shows) in three languages (English, Chinese, and Arabic) with structural information (syntax and predicate argument structure) and shallow semantics (word sense linked to an ontology and coreference). OntoNotes Release 4.0 is supported by the Defense Advance Research Project Agency, GALE Program Contract No. HR0011-06-C-0022.
OntoNotes Release 4.0 contains the content of earlier releases -- OntoNotes Release 1.0 LDC2007T21, OntoNotes Release 2.0 LDC2008T04 and OntoNotes Release 3.0 LDC2009T24 -- and adds newswire, broadcast news, broadcast conversation and web data in English and Chinese and newswire data in Arabic. This cumulative publication consists of 2.4 million words as follows: 300k words of Arabic newswire 250k words of Chinese newswire, 250k words of Chinese broadcast news, 150k words of Chinese broadcast conversation and 150k words of Chinese web text and 600k words of English newswire, 200k word of English broadcast news, 200k words of English broadcast conversation and 300k words of English web text.
The OntoNotes project builds on two time-tested resources, following the Penn Treebank for syntax and the Penn PropBank for predicate-argument structure. Its semantic representation will include word sense disambiguation for nouns and verbs, with each word sense connected to an ontology, and coreference. The current goals call for annotation of over a million words each of English and Chinese, and half a million words of Arabic over five years." (http://catalog.ldc.upenn.edu/LDC2011T03)
The OntoNotes corpus can be used to train the Name Finder. The corpus contains many different name types to train a model for a specific type only the built-in type filter option should be used.
The sample shows how to train a model to detect person names.
$ bin/opennlp TokenNameFinderTrainer.ontonotes -lang en -model en-ontonotes.bin \ -nameTypes person -ontoNotesDir ontonotes-release-4.0/data/files/data/english/ Indexing events using cutoff of 5 Computing event counts... done. 1953446 events Indexing... done. Sorting and merging events... done. Reduced 1953446 events to 1822037. Done indexing. Incorporating indexed data for training... done. Number of Event Tokens: 1822037 Number of Outcomes: 3 Number of Predicates: 298263 ...done. Computing model parameters ... Performing 100 iterations. 1: ... loglikelihood=-2146079.7808976253 0.976677625078963 2: ... loglikelihood=-195016.59754190338 0.976677625078963 ... cut lots of iterations ... 99: ... loglikelihood=-10269.902459614596 0.9987299367374374 100: ... loglikelihood=-10227.160010853702 0.9987314724850341 Writing name finder model ... done (2.315s) Wrote name finder model to path: /dev/opennlp/trunk/opennlp-tools/en-ontonotes.bin
Table of Contents
To explain what maximum entropy is, it will be simplest to quote from Manning and Schutze* (p. 589): “ Maximum entropy modeling is a framework for integrating information from many heterogeneous information sources for classification. The data for a classification problem is described as a (potentially large) number of features. These features can be quite complex and allow the experimenter to make use of prior knowledge about what types of informations are expected to be important for classification. Each feature corresponds to a constraint on the model. We then compute the maximum entropy model, the model with the maximum entropy of all the models that satisfy the constraints. This term may seem perverse, since we have spent most of the book trying to minimize the (cross) entropy of models, but the idea is that we do not want to go beyond the data. If we chose a model with less entropy, we would add `information' constraints to the model that are not justified by the empirical evidence available to us. Choosing the maximum entropy model is motivated by the desire to preserve as much uncertainty as possible. ”
So that gives a rough idea of what the maximum entropy framework is. Don't assume anything about your probability distribution other than what you have observed.
On the engineering level, using maxent is an excellent way of creating programs which perform very difficult classification tasks very well. For example, precision and recall figures for programs using maxent models have reached (or are) the state of the art on tasks like part of speech tagging, sentence detection, prepositional phrase attachment, and named entity recognition. On the engineering level, an added benefit is that the person creating a maxent model only needs to inform the training procedure of the event space, and need not worry about independence between features.
While the authors of this implementation of maximum entropy are generally interested using maxent models in natural language processing, the framework is certainly quite general and useful for a much wider variety of fields. In fact, maximum entropy modeling was originally developed for statistical physics.
For a very in-depth discussion of how maxent can be used in natural language processing, try reading Adwait Ratnaparkhi's dissertation. Also, check out Berger, Della Pietra, and Della Pietra's paper A Maximum Entropy Approach to Natural Language Processing, which provides an excellent introduction and discussion of the framework.
*Foundations of statistical natural language processing . Christopher D. Manning, Hinrich Schutze. Cambridge, Mass. : MIT Press, c1999.
We have tried to make the opennlp.maxent implementation easy to use. To create a model, one needs (of course) the training data, and then implementations of two interfaces in the opennlp.maxent package, EventStream and ContextGenerator. These have fairly simple specifications, and example implementations can be found in the OpenNLP Tools preprocessing components.
We have also set in place some interfaces and code to make it easier to automate the training and evaluation process (the Evalable interface and the TrainEval class). It is not necessary to use this functionality, but if you do you'll find it much easier to see how well your models are doing. The opennlp.grok.preprocess.namefind package is an example of a maximum entropy component which uses this functionality.
We have managed to use several techniques to reduce the size of the models when writing them to disk, which also means that reading in a model for use is much quicker than with less compact encodings of the model. This was especially important to us since we use many maxent models in the Grok library, and we wanted the start up time and the physical size of the library to be as minimal as possible. As of version 1.2.0, maxent has an io package which greatly simplifies the process of loading and saving models in different formats.
Table of Contents
The UIMA Integration wraps the OpenNLP components in UIMA Analysis Engines which can be used to automatically annotate text and train new OpenNLP models from annotated text.
The Cas Visual Debugger is shipped as part of the UIMA distribution and is a tool which can run the OpenNLP UIMA Annotators and display their analysis results. The source distribution comes with a script which can create a sample UIMA application. Which includes the sentence detector, tokenizer, pos tagger, chunker and name finders for English. This sample application is packaged in the pear format and must be installed with the pear installer before it can be run by CVD. Please consult the UIMA documentation for further information about the pear installer.
The OpenNLP UIMA pear file must be build manually. First download the source distribution, unzip it and go to the apache-opennlp/opennlp folder. Type "mvn install" to build everything. Now build the pear file, go to apache-opennlp/opennlp-uima and build it as shown below. Note the models will be downloaded from the old SourceForge repository and are not licensed under the AL 2.0.
$ ant -f createPear.xml Buildfile: createPear.xml createPear: [echo] ##### Creating OpenNlpTextAnalyzer pear ##### [copy] Copying 13 files to OpenNlpTextAnalyzer/desc [copy] Copying 1 file to OpenNlpTextAnalyzer/metadata [copy] Copying 1 file to OpenNlpTextAnalyzer/lib [copy] Copying 3 files to OpenNlpTextAnalyzer/lib [mkdir] Created dir: OpenNlpTextAnalyzer/models [get] Getting: http://opennlp.sourceforge.net/models-1.5/en-token.bin [get] To: OpenNlpTextAnalyzer/models/en-token.bin [get] Getting: http://opennlp.sourceforge.net/models-1.5/en-sent.bin [get] To: OpenNlpTextAnalyzer/models/en-sent.bin [get] Getting: http://opennlp.sourceforge.net/models-1.5/en-ner-date.bin [get] To: OpenNlpTextAnalyzer/models/en-ner-date.bin [get] Getting: http://opennlp.sourceforge.net/models-1.5/en-ner-location.bin [get] To: OpenNlpTextAnalyzer/models/en-ner-location.bin [get] Getting: http://opennlp.sourceforge.net/models-1.5/en-ner-money.bin [get] To: OpenNlpTextAnalyzer/models/en-ner-money.bin [get] Getting: http://opennlp.sourceforge.net/models-1.5/en-ner-organization.bin [get] To: OpenNlpTextAnalyzer/models/en-ner-organization.bin [get] Getting: http://opennlp.sourceforge.net/models-1.5/en-ner-percentage.bin [get] To: OpenNlpTextAnalyzer/models/en-ner-percentage.bin [get] Getting: http://opennlp.sourceforge.net/models-1.5/en-ner-person.bin [get] To: OpenNlpTextAnalyzer/models/en-ner-person.bin [get] Getting: http://opennlp.sourceforge.net/models-1.5/en-ner-time.bin [get] To: OpenNlpTextAnalyzer/models/en-ner-time.bin [get] Getting: http://opennlp.sourceforge.net/models-1.5/en-pos-maxent.bin [get] To: OpenNlpTextAnalyzer/models/en-pos-maxent.bin [get] Getting: http://opennlp.sourceforge.net/models-1.5/en-chunker.bin [get] To: OpenNlpTextAnalyzer/models/en-chunker.bin [zip] Building zip: OpenNlpTextAnalyzer.pear BUILD SUCCESSFUL Total time: 3 minutes 20 seconds
After the pear is installed start the Cas Visual Debugger shipped with the UIMA framework. And click on Tools -> Load AE. Then select the opennlp.uima.OpenNlpTextAnalyzer_pear.xml file in the file dialog. Now enter some text and start the analysis engine with "Run -> Run OpenNLPTextAnalyzer". Afterwards the results will be displayed. You should see sentences, tokens, chunks, pos tags and maybe some names. Remember the input text must be written in English.
For more information about how to use the integration please consult the javadoc of the individual Analysis Engines and checkout the included xml descriptors.
TODO: Extend this documentation with information about the individual components. If you want to contribute please contact us on the mailing list or comment on the jira issue OPENNLP-49.