Exploring the Maze of Multilingual Modeling

التفاصيل البيبلوغرافية
العنوان: Exploring the Maze of Multilingual Modeling
المؤلفون: Nezhad, Sina Bagheri, Agrawal, Ameeta
سنة النشر: 2023
المجموعة: Computer Science
مصطلحات موضوعية: Computer Science - Computation and Language, I.2.7
الوصف: Multilingual language models have gained significant attention in recent years, enabling the development of applications that meet diverse linguistic contexts. In this paper, we present a comprehensive evaluation of three popular multilingual language models: mBERT, XLM-R, and GPT-3. We assess their performance across a diverse set of languages, with a focus on understanding the impact of resource availability (general and model-specific), language family, script type, and word order on model performance, under two distinct tasks - text classification and text generation. Our findings reveal that while the amount of language-specific pretraining data plays a crucial role in model performance, we also identify other factors such as general resource availability, language family, and script type, as important features. We hope that our study contributes to a deeper understanding of multilingual language models to enhance their performance across languages and linguistic contexts.
نوع الوثيقة: Working Paper
URL الوصول: http://arxiv.org/abs/2310.05404
رقم الأكسشن: edsarx.2310.05404
قاعدة البيانات: arXiv