Abstract
This submission investigates alternative machine learning models for predicting the HTER score on the sentence level. Instead of directly predicting the HTER score, we suggest a model that jointly predicts the amount of the 4 distinct post-editing operations, which are then used to calculate the HTER score. This also gives the possibility to correct invalid (e.g. negative) predicted values prior to the calculation of the HTER score. Without any feature exploration, a multi-layer perceptron with 4 outputs yields small but significant improvements over the baseline.
Cite
CITATION STYLE
Avramidis, E. (2017). Sentence-level quality estimation by predicting HTER as a multi-component metric. In WMT 2017 - 2nd Conference on Machine Translation, Proceedings (pp. 534–539). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/w17-4758
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.