A novel mark-then-translate method, EasyProject, simplifies cross-lingual transfer learning for span-level NLP tasks by outperforming traditional word alignment methods while being easier to implement.
Pretraining decoder-only LLMs with active forgetting, a technique involving periodic resetting of token embeddings, enhances their cross-lingual transfer capabilities and allows for better adaptation to new languages without sacrificing performance in other languages.
INCLINE, a novel inference-time intervention framework, effectively bridges performance gaps between high-resource and low-resource languages in Large Language Models (LLMs) by aligning their internal representations, leading to significant performance improvements on various multilingual tasks without requiring costly retraining or fine-tuning.
Phonemic representations, particularly using the International Phonetic Alphabet (IPA), can mitigate performance gaps in cross-lingual transfer learning by reducing linguistic discrepancies between languages, especially for low-resource languages, as demonstrated by improved results in tasks like XNLI, NER, and POS tagging.
Transliteration-based post-training alignment improves cross-lingual transfer in multilingual language models, especially between related languages with different scripts, by aligning representations in the original and Latin scripts.
Establishing strong multilingual alignment before and during the pretraining of large language models significantly improves their ability to transfer knowledge and skills across languages, especially in low-resource scenarios.
Large language models can capture cross-lingual correspondence, which can be effectively elicited through self-translation to improve cross-lingual transfer performance.
Cross-lingual transfer learning can lead to catastrophic forgetting of previously acquired knowledge in the source language. This study compares different cross-lingual transfer strategies and fine-tuning approaches to measure and mitigate this effect.
The study demonstrates that language models can effectively transfer knowledge across diverse languages, with the transfer being largely independent of language proximity. This suggests the presence of language-agnostic representations that enable cross-lingual generalization.
다중 소스 크로스-언어 전이 학습을 위한 DA-Net의 효과적인 구조 소개