MELA is a comprehensive benchmark for linguistic acceptability judgment, covering 10 languages with 48K samples. It highlights the significance of in-language training data for accurate judgments and improved performance on syntax-related tasks. The study explores cross-lingual transfer and fine-tuning effects on Large Language Models (LLMs). Results show GPT-4 performing comparably to XLM-R, emphasizing the need for in-language training data. Probing experiments reveal enhanced syntax capacity acquisition through MELA training. The dataset aims to facilitate research on multilingual language models and syntactic competence acquisition.
In un'altra lingua
dal contenuto originale
arxiv.org
Approfondimenti chiave tratti da
by Ziyin Zhang,... alle arxiv.org 03-05-2024
https://arxiv.org/pdf/2311.09033.pdfDomande più approfondite