Pretrained large language models (LLMs) are versatile but suffer from format specialization during fine-tuning, reducing generalization. ProMoT proposes a two-stage fine-tuning approach to mitigate format specialization and enhance generalization. Experimental results show improved performance on diverse tasks with ProMoT compared to standard fine-tuning methods.
To Another Language
from source content
arxiv.org
Key Insights Distilled From
by Yihan Wang,S... at arxiv.org 03-14-2024
https://arxiv.org/pdf/2211.00635.pdfDeeper Inquiries