Alibaba Unveils Babel: A Multilingual Language Model on Hugging Face
Babel-9B is optimized for efficient single-GPU inference and fine-tuning, while Babel-83B sets a new standard for open multilingual LLMs. The models have been extensively evaluated on multilingual tasks, demonstrating superior performance compared to other open LLMs of similar size. Babel-9B-Chat leads among 10B-sized LLMs, and Babel-83B-Chat performs comparably to GPT-4o on certain tasks.
The models employ a unique layer extension technique to expand capacity without compromising performance, focusing on languages that are often underrepresented in existing models. This approach allows Babel to improve accuracy in languages with limited training data, achieving up to 10% better performance in these languages compared to previous multilingual LLMs.
We hope you enjoyed this article.
Consider subscribing to one of several newsletters we publish. For example, in the Daily AI Brief you can read the most up to date AI news round-up 6 days per week.
Also, consider following our LinkedIn page AI Brief.
Subscribe to Daily AI Brief
Daily report covering major AI developments and industry news, with both top stories and complete market updates