A Comparative Review of Multilingual Language Models: Evaluation of mBERT and XLM-R on Cross-Lingual Tasks

Publish Year: 1404
نوع سند: مقاله کنفرانسی
زبان: English
View: 43

This Paper With 8 Page And PDF Format Ready To Download

  • Certificate
  • من نویسنده این مقاله هستم

استخراج به نرم افزارهای پژوهشی:

لینک ثابت به این Paper:

شناسه ملی سند علمی:

ICPCONF11_137

تاریخ نمایه سازی: 1 آذر 1404

Abstract:

Multilingual pre-trained models have revolutionized natural language processing (NLP) by enabling cross-language transfer and enhancing performance in low-resource languages. This review article examines three fundamental studies on multilingual language models: the XTREME benchmark for evaluating cross-language generalization, the XLM-R architecture with its large-scale training, and a comparative evaluation of multilingual advertisement recognition using mBERT, XLM-R, and mT۵. We summarize the methodologies, datasets, evaluation tasks, and results from these studies, and provide a hybrid comparison that highlights the strengths and limitations of mBERT and XLM-R across various NLP tasks.

Authors

Mohammad Taheri Asl

PhD student in department of Computer Engineering, Islamic Azad University, Qom

Ahmad Sharif

Faculty member of department of Computer engineering, Qo.C., Islamic Azad University, Qom, Iran