Skip to content

Latest commit

 

History

History
64 lines (40 loc) · 2.92 KB

Pretrained_Models.md

File metadata and controls

64 lines (40 loc) · 2.92 KB

Pretrained Sequence Tagging Models

In the following some pre-trained models are provided for different common sequence tagging tasks. These models can be used by executing:

python RunModel.py modelname.h5 input.txt

For the English models, we used the word embeddings by Komninos et al.. For the German files, we used the word embeddings by Reimers et al..

POS

We trained POS-tagger on the Universal Dependencies v1.3 dataset: Trained on universal dependencies v1.3 Englisch:

Language Development (Accuracy) Test (Accuracy)
English (UD) 95.58% 95.58%
German (UD) 94.50% 93.88%

Further, we trained models on the Wall Street Journal:

Language Development (Accuracy) Test (Accuracy)
English (WSJ) 97.33% 97.39%

The depicted performance is accuracy.

Chunking

Trained on CoNLL 2000 Chunking dataset. Performance is F1-score.

Language Development (F1) Test(F1)
English (CoNLL 2003) 95.30% 94.71%

NER

Trained on CoNLL 2003 and GermEval 2014

Language Development (F1) Test (F1)
English (CoNLL 2003) 93.87% 90.22%
German (CoNLL 2003) 81.15% 77.70%
German (GermEval 2014) 80.93% 78.94%

Entities

Trained on ACE 2005 (https://catalog.ldc.upenn.edu/LDC2006T06)

Language Development (F1) Test (F1)
English 83.93% 85.68%

Events

Trained on TempEval3 (https://www.cs.york.ac.uk/semeval-2013/task1/)

Language Development (F1) Test (F1)
English - 83.45%