TigerLLM -- A Family of Bangla Large Language Models
TigerLLM -- A Family of Bangla Large Language Models
The development of Large Language Models (LLMs) remains heavily skewed towards English and a few other high-resource languages. This linguistic disparity is particularly evident for Bangla - the 5th most spoken language. A few initiatives attempted to create open-source Bangla LLMs with performance still behind high-resource languages and limited reproducibility. To address this gap, we introduce TigerLLM - a family of Bangla LLMs. Our results demonstrate that these models surpass all open-source alternatives and also outperform larger proprietary models like GPT3.5 across standard benchmarks, establishing TigerLLM as the new baseline for future Bangla language modeling.
Marcos Zampieri、Nishat Raihan
印欧语系
Marcos Zampieri,Nishat Raihan.TigerLLM -- A Family of Bangla Large Language Models[EB/OL].(2025-03-13)[2025-04-29].https://arxiv.org/abs/2503.10995.点此复制
评论