A Comparative Review of Multilingual Language Models: Evaluation of mBERT and XLM-R on Cross-Lingual Tasks
Publish Year: 1404
نوع سند: مقاله کنفرانسی
زبان: English
View: 43
This Paper With 8 Page And PDF Format Ready To Download
- Certificate
- من نویسنده این مقاله هستم
استخراج به نرم افزارهای پژوهشی:
شناسه ملی سند علمی:
ICPCONF11_137
تاریخ نمایه سازی: 1 آذر 1404
Abstract:
Multilingual pre-trained models have revolutionized natural language processing (NLP) by enabling cross-language transfer and enhancing performance in low-resource languages. This review article examines three fundamental studies on multilingual language models: the XTREME benchmark for evaluating cross-language generalization, the XLM-R architecture with its large-scale training, and a comparative evaluation of multilingual advertisement recognition using mBERT, XLM-R, and mT۵. We summarize the methodologies, datasets, evaluation tasks, and results from these studies, and provide a hybrid comparison that highlights the strengths and limitations of mBERT and XLM-R across various NLP tasks.
Keywords:
Authors
Mohammad Taheri Asl
PhD student in department of Computer Engineering, Islamic Azad University, Qom
Ahmad Sharif
Faculty member of department of Computer engineering, Qo.C., Islamic Azad University, Qom, Iran