Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Unbabel's Participation in the WMT19 Translation Quality Estimation Shared Task (1907.10352v2)

Published 24 Jul 2019 in cs.CL

Abstract: We present the contribution of the Unbabel team to the WMT 2019 Shared Task on Quality Estimation. We participated on the word, sentence, and document-level tracks, encompassing 3 language pairs: English-German, English-Russian, and English-French. Our submissions build upon the recent OpenKiwi framework: we combine linear, neural, and predictor-estimator systems with new transfer learning approaches using BERT and XLM pre-trained models. We compare systems individually and propose new ensemble techniques for word and sentence-level predictions. We also propose a simple technique for converting word labels into document-level predictions. Overall, our submitted systems achieve the best results on all tracks and language pairs by a considerable margin.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (8)
  1. Fabio Kepler (2 papers)
  2. Jonay Trénous (3 papers)
  3. Marcos Treviso (17 papers)
  4. Miguel Vera (2 papers)
  5. António Góis (7 papers)
  6. M. Amin Farajian (3 papers)
  7. António V. Lopes (4 papers)
  8. André F. T. Martins (113 papers)
Citations (58)