To evaluate machine translation quality across several genres and
languages we provide a collection of benchmark data sets with genre
annotations. If you use this data set, please cite the following paper:
- Marlies van der Wees, Arianna Bisazza, and Christof Monz. Evaluation of Machine Translation Performance Across Multiple Genres and Languages. In Proceedings of the 11th edition of the Language Resources and Evaluation Conference (LREC-2018).
The dataset is available here.