1 |
Towards Explainable Evaluation Metrics for Natural Language Generation ...
|
|
|
|
BASE
|
|
Show details
|
|
2 |
Pushing the right buttons: adversarial evaluation of quality estimation
|
|
|
|
In: Proceedings of the Sixth Conference on Machine Translation ; 625 ; 638 (2022)
|
|
BASE
|
|
Show details
|
|
5 |
Continual Quality Estimation with Online Bayesian Meta-Learning ...
|
|
|
|
BASE
|
|
Show details
|
|
7 |
Findings of the WMT 2021 Shared Task on Quality Estimation ...
|
|
|
|
BASE
|
|
Show details
|
|
8 |
Pushing the Right Buttons: Adversarial Evaluation of Quality Estimation ...
|
|
|
|
BASE
|
|
Show details
|
|
10 |
deepQuest-py: large and distilled models for quality estimation
|
|
|
|
BASE
|
|
Show details
|
|
11 |
Findings of the WMT 2021 shared task on quality estimation
|
|
|
|
In: 689 ; 730 (2021)
|
|
BASE
|
|
Show details
|
|
12 |
deepQuest-py: large and distilled models for quality estimation
|
|
|
|
In: Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing: System Demonstrations ; 382 ; 389 (2021)
|
|
BASE
|
|
Show details
|
|
13 |
Backtranslation feedback improves user confidence in MT, not quality
|
|
|
|
BASE
|
|
Show details
|
|
14 |
Knowledge distillation for quality estimation
|
|
|
|
In: 5091 ; 5099 (2021)
|
|
BASE
|
|
Show details
|
|
15 |
MLQE-PE: A Multilingual Quality Estimation and Post-Editing Dataset ...
|
|
|
|
BASE
|
|
Show details
|
|
16 |
Unsupervised quality estimation for neural machine translation
|
|
|
|
In: 8 ; 539 ; 555 (2020)
|
|
BASE
|
|
Show details
|
|
17 |
An exploratory study on multilingual quality estimation
|
|
|
|
In: 366 ; 377 (2020)
|
|
Abstract:
This is an accepted manuscript of an article published by ACL, available online at: https://www.aclweb.org/anthology/2020.aacl-main.39 The accepted version of the publication may differ from the final published version. ; Predicting the quality of machine translation has traditionally been addressed with language-specific models, under the assumption that the quality label distribution or linguistic features exhibit traits that are not shared across languages. An obvious disadvantage of this approach is the need for labelled data for each given language pair. We challenge this assumption by exploring different approaches to multilingual Quality Estimation (QE), including using scores from translation models. We show that these outperform singlelanguage models, particularly in less balanced quality label distributions and low-resource settings. In the extreme case of zero-shot QE, we show that it is possible to accurately predict quality for any given new language from models trained on other languages. Our findings indicate that state-of-the-art neural QE models based on powerful pre-trained representations generalise well across languages, making them more applicable in real-world settings.
|
|
Keyword:
machine translation; multilingual; multitask learning; quality estimation; zero-shot learning
|
|
URL: http://hdl.handle.net/2436/623698
|
|
BASE
|
|
Hide details
|
|
18 |
BERGAMOT-LATTE submissions for the WMT20 quality estimation shared task
|
|
|
|
In: 1010 ; 1017 (2020)
|
|
BASE
|
|
Show details
|
|
19 |
Findings of the WMT 2020 shared task on quality estimation
|
|
|
|
In: 743 ; 764 (2020)
|
|
BASE
|
|
Show details
|
|
20 |
MLQE-PE: A multilingual quality estimation and post-editing dataset
|
|
|
|
BASE
|
|
Show details
|
|
|
|