This survey presents an overview of the evolution of Multilingual Large Language Models (MLLMs), tracing their development from monolingual Large Language Models (LLMs). It highlights the key techniques that contribute to the success of MLLMs, including transformer architecture, pre-training techniques, and reinforcement learning with human feedback.
The survey then explores the multilingual capacities of MLLMs, discussing the challenges they face due to language imbalance in training corpora and the potential for cross-lingual transfer learning. It provides an in-depth analysis of the widely utilized multilingual corpora and datasets for training and fine-tuning MLLMs, offering insights into their language distribution, data sources, and language coverage.
Next, the survey delves into the topic of multilingual representation alignment, categorizing the existing approaches into static, contextual, and combined methods. It examines the factors that affect the performance of these alignment techniques, such as initial alignment solution, language typological distance, and pre-training data and settings of MLLMs.
Finally, the survey discusses the issue of bias in MLLMs, addressing questions about the types of bias present, the available debiasing techniques, and the impact of bias removal on model performance. It also summarizes the existing bias evaluation datasets for MLLMs.
Throughout the survey, the author aims to facilitate a deeper understanding of MLLMs and their potential in various domains, while also highlighting the critical challenges and promising research directions in this field.
Naar een andere taal
vanuit de broninhoud
arxiv.org
Belangrijkste Inzichten Gedestilleerd Uit
by Yuemei Xu,Li... om arxiv.org 04-02-2024
https://arxiv.org/pdf/2404.00929.pdfDiepere vragen