This survey provides a comprehensive overview of techniques to improve cross-lingual alignment in multilingual language models. The authors first define two main views of cross-lingual alignment - the "similarity-based" view and the "subspace-based" view. They then present a taxonomy of methods for improving alignment, categorized by their initialization (from existing model or from scratch) and data requirements (parallel data at sentence or word level, target task data, or other sources).
The authors discuss the effectiveness of contrastive training for cross-lingual transfer, noting that pre-training alone does not determine performance. They also highlight that related languages tend to be more aligned within the models. The authors argue that "strong" alignment, as defined by the similarity-based view, may not be necessary for all tasks, and that effectively trading off language-neutral and language-specific information is key.
The survey then discusses the emerging challenges posed by multilingual generative models, where simply maximizing cross-lingual alignment can lead to wrong-language generation. The authors call for future methods that can balance cross-lingual semantic information with language-specific factors, enabling fluent and relevant generation in multiple languages.
In un'altra lingua
dal contenuto originale
arxiv.org
Approfondimenti chiave tratti da
by Kath... alle arxiv.org 04-10-2024
https://arxiv.org/pdf/2404.06228.pdfDomande più approfondite