Meta Unveils Machine Translation for Over 1,600 Languages
- •Meta’s OMT system scales machine translation from 200 to over 1,600 languages.
- •Specialized 1B-8B parameter models outperform 70B LLM baselines in translation quality.
- •New evaluation tools like BOUQuET ensure reliable testing across diverse, underrepresented linguistic families.
Meta AI researchers have achieved a massive leap in linguistic inclusivity with the introduction of Omnilingual Machine Translation (OMT). While previous state-of-the-art systems like No Language Left Behind (NLLB) supported 200 languages, OMT expands this coverage to over 1,600. This breakthrough specifically targets the "generation bottleneck," a common issue where AI models can understand many dialects but struggle to write or speak them accurately.
The system utilizes two distinct architectural approaches: OMT-LLaMA, which adapts a decoder-only model for translation using retrieval-augmented techniques, and OMT-NLLB, an encoder–decoder framework. By integrating massive multilingual corpora with synthetic data and manually curated bitext (translated sentence pairs), the team has enabled coherent generation for thousands of previously marginalized or endangered languages.
Remarkably, the study finds that smaller, specialized models ranging from 1 billion to 8 billion parameters can actually exceed the performance of much larger 70 billion parameter general-purpose models. This efficiency suggests that focused training on translation-specific data is more effective than sheer model size. To support the broader research community, Meta has released the BOUQuET evaluation dataset, currently the largest collection of its kind for testing multilingual fidelity across wide linguistic families.