Translation Quality Report. December 2023

The goal of this report is to compare translation quality between old and new language models. New models have not only improved quality but performance and memory usage. We used the BLEU metric and primarily Flores 101 test set in the report.

BLEU is the most popular metrics in the world for machine translation evaluation. Flores 101 test set was released by Facebook Research and has the biggest language pair coverage.

Quality metrics description

BLEU

BLEU is an automatic metric based on n-grams. It measures the precision of n-grams of the machine translation output compared to the reference, weighted by a brevity penalty to punish overly short translations. We use a particular implementation of BLEU, called sacreBLEU. It outputs corpus scores, not segment scores.

References

  • Papineni, Kishore, S. Roukos, T. Ward and Wei-Jing Zhu. “Bleu: a Method for Automatic Evaluation of Machine Translation.” ACL (2002).
  • Post, Matt. “A Call for Clarity in Reporting BLEU Scores.” WMT (2018).

Improved Language Models

BLEU Metrics

BLEU Metrics

Improved Language Models. December 2023

Language pairs

Note: The lower size of models on the hard drive means the lower consumption of GPU memory which leads to decreased deployment costs. Lower model size has better performance in translation time. The approximate usage of GPU memory is calculated as hard drive model size x 1.2

Conclusion

BLEU stands for Bilingual Evaluation Understudy - a wide spread metric, used for assessing machine translation quality. The present report compares the BLEU scores of the old and new language models with various language pairs. Its major findings are that the new models have higher BLEU scores demonstrating better translation quality. The report also shows improvements in memory usage which leads to lower GPU memory consumption and helps to reduce deployment costs.


Frequently Asked Questions (FAQ)

What is the blue score BiLingual evaluation understudy?

It’s an algorithm for evaluating the quality of a machine translated text. The BLEU score is a number between 0 and 1 that measures the similarity of the machine-translated text to a set of high quality reference translations.

What does BLEU measure?

BLEU is a metric for evaluating a generated phrase to a reference phrase. A perfect match is a score of 1.0, and a perfect mismatch represents a score of 0.0.

What is the metric of NLP BLEU?

It is a metric used in natural language processing (NLP) and machine translation to evaluate the quality of generated text in comparison with one or more high quality reference translations.

More fascinating reads await

Speech Recognition Quality Comparison

Speech Recognition Quality Comparison

April 30, 2025

Machine Translation in the Military Sphere

Machine Translation in the Military Sphere

April 16, 2025

The Best English-to-Arabic Translation Model in the World

The Best English-to-Arabic Translation Model in the World

March 6, 2025

Contact Us

* Required fields

By submitting this form, I agree that the Terms of Service and Privacy Policy will govern the use of services I receive and personal data I provide respectively.

Email

Completed

Your request has been sent successfully

×