MELA is a comprehensive benchmark for linguistic acceptability judgment, covering 10 languages with 48K samples. It highlights the significance of in-language training data for accurate judgments and improved performance on syntax-related tasks. The study explores cross-lingual transfer and fine-tuning effects on Large Language Models (LLMs). Results show GPT-4 performing comparably to XLM-R, emphasizing the need for in-language training data. Probing experiments reveal enhanced syntax capacity acquisition through MELA training. The dataset aims to facilitate research on multilingual language models and syntactic competence acquisition.
In eine andere Sprache
aus dem Quellinhalt
arxiv.org
Wichtige Erkenntnisse aus
by Ziyin Zhang,... um arxiv.org 03-05-2024
https://arxiv.org/pdf/2311.09033.pdfTiefere Fragen