FuxiMT: Sparsifying Large Language Models for Chinese-Centric Multilingual Machine Translation
Shaolin Zhu, Tianyu Dong, Bo Li, Deyi Xiong
2025-05-26
Summary
This paper talks about FuxiMT, a new machine translation system that focuses on translating between Chinese and many other languages, using a special technique to make the language model more efficient and powerful.
What's the problem?
The problem is that most translation models struggle with languages that don't have a lot of data, and it's especially hard to get good translations when going to or from Chinese in these low-resource situations. Regular models can also be slow or require a lot of computer power.
What's the solution?
The researchers built FuxiMT by making the language model 'sparse,' which means it only uses the parts it needs for each translation, making it faster and more efficient. This model is centered on Chinese and can handle translations with 65 different languages, even doing well when it hasn't seen much data for some of them.
Why it matters?
This is important because it helps people communicate better across many languages, especially those that don't have a lot of digital resources, and it makes translation tools more accessible and effective for everyone.
Abstract
FuxiMT, a Chinese-centric multilingual machine translation model utilizing a sparsified large language model, demonstrates superior performance in low-resource scenarios and strong zero-shot capabilities across 65 languages.