Google Wants to Improve Human Translation Evaluation with This Simple Step
slator.com·16h
Flag this post

As AI translation quality becomes better and better, a crucial question has emerged: can human evaluation keep up? A new study from Google researchers argues that it can and explains how.

In their October 28, 2025, paper, researchers Parker Riley, Daniel Deutsch, Mara Finkelstein, Colten DiIanni, Juraj Juraska, and Markus Freitag proposed a refinement to the Multidimensional Quality Metrics (MQM) framework: re-annotation.

Rather than relying on a single pass, a second human rater — either the same or a different one — reviews an existing annotation — whether human- or machine-generated — correcting, deleting, or adding error spans.

The …

Similar Posts

Loading similar posts...