Appraise: An Open-Source Toolkit for Manual Evaluation of Machine Translation Output

Christian Federmann

In: The Prague Bulletin of Mathematical Linguistics (PBML) 98 Pages 25-35 Charles University in Prague Prague, Czech Republic 9/2012.


We describe Appraise, an open-source toolkit supporting manual evaluation of machine translation output. The system allows to collect human judgments on translation output, implementing annotation tasks such as 1) quality checking, 2) translation ranking, 3) error classification, and 4) manual post-editing. It features an extensible, XML-based format for import/ export and can easily be adapted to new annotation tasks. The current version of Appraise also includes automatic computation of inter-annotator agreements allowing quick access to evaluation results. Appraise is actively developed and used in several MT projects.

Weitere Links

German Research Center for Artificial Intelligence
Deutsches Forschungszentrum für Künstliche Intelligenz